...

/

Stochastic Gradient Descent

Stochastic Gradient Descent

Learn about SGD-based optimizers in JAX and Flax.

We'll cover the following...

SGD implements stochastic gradient descent with support for momentum and Nesterov accelerationNesterov acceleration refers to a method of accelerating the convergence of iterative optimization algorithms commonly used in machine learning.. Momentum makes obtaining optimal model weights faster by accelerating gradient descent in a certain direction.

Press + to interact
Gradient function
Gradient function

Let’s ...

Access this course and 1400+ top-rated courses and projects.