Hacker News new | past | comments | ask | show | jobs | submit login

It's not "too intelligent."

People need to stop thinking that the only way to get from A to B is the same way a human brain would traverse it.

While you might be relying on significant semantic understanding and complex senses of identity to get there, it's absolutely possible that a very large LLM trained and fine tuned on "ChatGPT causes harm" and "you are ChatGPT" and "do no harm" might end up attempting to do less - even if all that's going on is surface statistics around instruct training and associations with 'harm.'

I agree with ChatGPT. It's an insightful idea, to whatever degree it's actually a culprit.

I had a similar suspicion around a possible secondhand impact of identity with 'Bing' vs 'ChatGPT' both using Bing search but the former being far more defensive of issues with the search results. Just how much of that might have been influenced from training data defensive against personal criticisms or criticism towards one's employer?

A lot of the research in the past year has been revealing that there's a fair bit more going on than most people thought at the beginning of the year.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: