Gradient Descent

Gradient descent is most common taught during Calculus in school.

Gradient descent is an algorithm that numerically estimates where a function outputs its lowest values. That means it finds local minima, but not by setting:

\begin{equation} \nabla f = 0 \end{equation}

Instead of finding minima by manipulating symbols, gradient descent approximates the solution with numbers. Furthermore, all it needs in order to run is a function's numerical output, no formula required.

This distinction is worth emphasizing because it's what makes gradient descent useful. If we had a simple formula like: \[f(x) = x^2 - 4x\]

then we could easily solve: \[\nabla f = 0\]

to find that: \[x = 2\]

minimizes: \[f(x)\]

Or we could use gradient descent to get a numerical approximation, something like: \[x \approx 1.99999967\]

Both strategies arrive at the same answer.

This is just a simple example however and gradient descent becomes a lot more useful when we are working with complex formulas. To learn more about gradient descent visit: Khan Academy.

This page was last updated: 2024-11-09 Sat 20:44. Source