GradientDescent
SGD

What are the differences between gradient descent and stochastic gradient descent? When would you use one over the other?

machine learning
Junior Level

Gradient descent and stochastic gradient descent (SGD) are optimization algorithms used to minimize a function, typically associated with minimizing the error in a model.

The primary differences between the two are the following:

Gradient Descent (GD)

  • In standard...

Code Labs Academy © 2024 All rights reserved.