Question: In Distributed Data Parallel (DDP) training, what is synchronizing across GPUs? Question 19 options: a) Only the gradients b) Only the model parameters c) Both
In Distributed Data Parallel (DDP) training, what is synchronizing across GPUs? Question 19 options: a) Only the gradients b) Only the model parameters c) Both gradients and model parameters d) Neither gradients nor parameters
Step by Step Solution
There are 3 Steps involved in it
Get step-by-step solutions from verified subject matter experts
