Multivariate Calculus
This lesson will introduce multivariate calculus in JAX.
We'll cover the following...
We can build on our linear algebra refresher to use auto differentiation for vectors and matrices as well. These are instrumental in neural networks when we apply gradient-based optimization on whole matrices.
Gradient
If we have a multivariate function (a function of more than one variable), , we calculate its derivative by taking the partial derivative with respect to every (input/independent) variable.
For example, if we have:
it’s derivative, represented by ∇, will be calculated as:
...