Hacker News new | past | comments | ask | show | jobs | submit login

It's very cool work.

I've been wondering for a while at what the next steps in adding 'inefficiencies' to AI processing would look like, commenting the other day to a friend that what's needed in the next 18 months is getting AI to be able to replicate the Eureka moments in the shower where latent information is reconstructed in parallel to processing tangential topics.

Going from "attention is all you need" to "attention and curiosity is what you need" seems like a great next step!




> getting AI to be able to replicate the Eureka moments in the shower where latent information is reconstructed in parallel to processing tangential topics.

I've been playing with this part specifically and it's really amazing stuff.

Having the model concurrently model internal monologue and output for a task, but allowing the internal monologue to be as focused or unfocused as the model sees fit.

You end up with situations where you have it working on a naming task for example, and the model starts imagining the warmth of a coffee cup on the desk, or traffic building up outside for a future appointment with a non-existent person, and then returns back to the task at hand with non-obvious tangents that it'd probably never have uncovered if it was only predicting on tokens related to the original goal of naming something.

It gets even more interesting when you inject variability into the process via the API (for example, telling it to use certain letters pulled from an RNG inside the next iteration of internal monologue).


This sounds fascinating. How do you make an internal monologue? Do you have any reading or samples to look at? Sorry for the ignorance, I’m a dev but not in the AI space.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: