Hacker News new | past | comments | ask | show | jobs | submit login

Specifically for neural networks, is there any alternative for backpropagation and gradient descent which guarantee finding the global minimum?



Unlikely given the dimensionality and complexity of the search space. Besides, we probably don’t even care about the global minimum: the loss we’re optimising is a proxy for what we really care about (performance on unseen data). Counter-example: a model that perfectly memorises the training data can be globally optimal (ignoring regularization), but is not very useful.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: