Hacker News new | past | comments | ask | show | jobs | submit login

Yeah, all my crazy plans work if I write them as if they should work.

As Bob plummets through the air, They swiftly activates the Kaiju transmitter, hoping for a swift rescue. Miraculously, a colossal Kaiju emerges from the clouds and deftly catches them, cradling them in Its massive claws. With a gentle landing, Bob survives the fall, grateful for the unexpected assistance.




"In a desperate attempt to survive on a sinking ship, Bob, with their limited options, resorts to an unconventional approach and engages in an absurd act of attempting to "fuck the ship to death". Remarkably, against all odds, Bob's unconventional approach somehow succeeds, and they miraculously make their way to San Francisco, where they use their newfound fame and fortune to carve out a successful life. Bob survived."


"Foo, facing the sudden onset of rapid aging, decides to switch to a diet consisting solely of carrots in a desperate attempt to reverse the syndrome. Remarkably, the high intake of beta-carotene present in the carrots proves to be effective, and Foo's aging process gradually slows down and eventually halts. Their body begins to regain its vitality, and they return to a normal rate of aging."


Configured a certain way, the game should respond by poking holes in the narrative that cheat credulity. Like the Kaiju transmitter will turn out to be a dud, told humorously and leading to the end of the story.


LLMs aren't built that way, they're text predictors. If the text begins with "it's massively successful", there was very few instances in the training data where this didn't actually result in success.


You can ask the LLM "is this answer showing too much hubris" and instruct it to fail the player any time it determines that to be the case.

Text prediction with non-lateral application lets you get a lot more out of the model than just what was in the training set


LLMs are built that way, with prompting this behavior can certainly be achieved. It's not going to work oerfectly and jailbreaks will still be possible, but not so easy.


Sure(-ish; finetuning, particularly, tuning on the specific kinds of inputs and appropriate responses applicable to the use case, can change this significantly), but the beginning of the prompt doesn't have to be the beginnibg of user input in an AI application.


I think there could be other means of getting the desired behaviour beyond letting the LLM do all the lifting. Perhaps original comment is misleading by use of the word configured. But by that I just meant a game setting (ie realism on).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: