Hacker News new | past | comments | ask | show | jobs | submit login

I believe GPT-4 pre-rlhf was much more accurate overall.

If you finetune it in formats that humans likes, it actually gains similar biases.

I believe the ‘sparks of AGI’ talks about this, where models can much more accurately predict the probability of events than humans.

After RLHF, it mimicks human bias. So it might be that we can create these models, but we just don’t like them/like to use them.




RLHF = nerfing and censorship

What, did you think the plebs would get access to the real deal?




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: