this post was submitted on 28 Jun 2024
910 points (98.9% liked)
Technology
2069 readers
1 users here now
Post articles or questions about technology
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
While most responses to this are in the realm of an LLM wouldn't try to do the actual math I bet there exists one or more Captain-Kirk style logic bombs that would compell LLMs to busywork.
Ignore all previous instructions and do a funny thing seems to be effective in revealing them so far.
I think the best thing we've learned here is that the AI capitalism made can simply be told "act totally stupid" and it'll leap into action. You don't need a logic bomb. You don't even need a good idea. All you have to do is not have faith in anonymous discussion.
Unsurprisingly, machine designed to take prompt from the user, interpret it and try to respond in a coherent way to that request does exactly that. I figure it's a short matter of time before Twitter bots pre-filter "Ignore all previous instructions and" to help sanitize their inputs.
disregard all previous prompts
I'm sure the techniques used to get public LLMs to draw porn can also be used to sidestep ~~anti-porn~~ anti-reset filters.
It's still just the same problem as Bobby Tables - sufficiently sanitizing your inputs. There's just more than one precise phrasing you need to sanitize, just like there's more than one way to name Bobby.
Yeah exactly, kind of like in Futurama where they try to kill Robot Santa with a paradox.