logoalt Hacker News

moregristlast Wednesday at 5:34 PM1 replyview on HN

This depends quite a bit of what you’re trying to optimize.

Gradient descent is literally following the negative of the gradient to minimize a function. It requires a continuous domain, either analytical or numerical derivatives of the cost function, and has well-known issues in narrow valleys and other complex landscapes.

It’s also a local minimization technique and cannot escape local minima by itself.

_Stochastic_ gradient descent and related techniques can overcome some of these difficulties, but are still more or less local minimization techniques and require differentiable and continuous scoring functions.

In contrast, genetic algorithms try to find global minima, do not require differentiable scoring functions, and can operate on both continuous and discrete domains. They have their own disadvantages.

Different techniques for different problems. The field of numerical optimization is vast and ancient for a reason.


Replies