Hacker News new | past | comments | ask | show | jobs | submit login

I think this is an underappreciated perspective. The simplest model of a reasoning process, at scale, is the reasoning process itself! That said, I haven't come across any research directly testing that hypothesis with transformers. Do you know of any?

The closest I've seen is a paper on OthelloGPT using linear probes to show that it does in fact learn a predictive model of Othello board states (which can be manipulated at inference time, so it's causal on the model's behaviour).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: