Hacker News new | past | comments | ask | show | jobs | submit login
LLMs are not even good wordcels (ferrei.ro)
3 points by PaulHoule 27 days ago | hide | past | favorite | 3 comments



Current LLMs are not aware of the letters in text. It's too expensive computationally to assign a token to each letter. They operate at the level of tokens, which tend to be word-level: each morph-em-e is broke-n down in-to seg-ment-s and assign-ed an int-e-ger.

The result is that LLMs are quite oblivious to spelling and such. There are some exceptions working at the letter level, but mostly, yeah. This applies to ChatGPT etc.


Most LLMs can't even talk about the tokenization they use which I think indicates that they don't have the right knowledge to do wordplay at the letter level.


> One thing that is clear to me is that we’re vastly overestimating the “intelligence” of LLMs, if it even makes sense to use that word in this context. We’re being fooled by their apparent capacity to produce human-sounding text.

Who is this "we"?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: