TheBoohers.org

Leader | Engineer | Hacker | Dad

Gradient Descent

It has been awhile since I’ve studied optimization, but gradient descent is always good to brush up on. Most optimization involves derivatives. Often known as the method of steepest descent, gradient descent works by taking steps proportional to the negative of the gradient of the function at the current point. Mathematically, gradient descent is defined […]

Read More