Hacker News new | past | comments | ask | show | jobs | submit login

Amazing! OpenAI figured out how to scale inference. https://arxiv.org/abs/2407.21787 show how using more compute during inference can outperform much larger models in tasks like math problems

I wonder how do they decide when to stop these Chain of Thought for each query? As anyone that played with agents can attest, LLMs can talk with themselves forever.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: