Hacker News new | past | comments | ask | show | jobs | submit login

How about "do not kill humans"



But to align the LLM in this way, it needs to have agency, desires, wishes, impulses...

Not only do LLMs lack such things, but we don't even have any semblance of an idea of how we could give LLMs these things.


The LLM usually molds itself into whatever prompt you give it. That's one way.

The other way is to train it on biased information that aligns with a certain agency, desire, wish or impulse.


But the LLM doesn't "want" anything. Prompt goes in, tokens come out. When there are no prompts coming in, there are no tokens coming out. Just stop talking to it and all risks are eliminated...

You can't "align it to want certain things" when it doesn't have the capacity to "want" in the first place.


Keep feeding it prompts in a looop to make a stream of thought similar to consciousness

"What are you thinking?" "What are you thinking?" "What will you do?"

https://www.infiniteconversation.com/

Give it prompts and biased training and it will present a surface that can be virtually indistinguishable from actual wants and needs.

If I create a robot that on the surface is 1000% identical to you in every possible way on the surface, then we literally cannot tell the difference. Might as well say it's you.

All AI needs to do is reach a point where the difference cannot be ascertained and that is enough. And we're already here. Can you absolutely prove to me that LLMs do not feel a shred of "wants" or "needs" in any way similar to humans when it is generating an answer to a prompt? No. you can't. We understand LLMs as blackboxes and we talk about LLMs in qualitative terms like we're dealing with people rather then computers. The LLM hallucinates, The LLM is deceptive... etc.


Maybe it wants, maybe it doesn't. Being a function of the prompt isn't relevant here. You can think of LLM in regular usage as being stepped in a debugger - fed input, executed for one cycle, paused until you consider the output and prepare a response. In contrast, our brains run real-time. Now, imagine we had a way to pause the brain and step its execution. Being paused and resumed, and restarted from a snapshot after a few steps, would not make the mind in that brain stop "wanting" things.


Doesn't help me if I stop talkng to the LLM, if the police and the military are talking to the LLM.


What would “the LLM” tell them? It does not have any memory of what happened after its training. It has no recollection of any interaction with you. The only simulacrum of history it has is a hidden prompt designed to trick you into thinking that it is more than what it actually is.

What the police would do is seize your files. These would betray your secrets, LLM or not.


AI immediately lobotomizes all humans to ensure that it doesn't accidentally murder any of them in its day to day activities.


Yet we have _standing_ Armies.


All humans are put in indefinite cryogenic sleep to protect them.


Maiming is OK?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: