Hacker News new | past | comments | ask | show | jobs | submit login

Gradient descent can also jump out of a local minimum if the learning rate is large enough, so they're equal in that sense.

But it does make sense that a random walk would be more efficient in very high dimensional problems!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: