Another day, another preprint paper shocked that it’s trivial to make a chatbot spew out undesirable and horrible content. [arXiv] How do you break LLM security with “prompt injection”?…
Look, AI will be perfect as soon as we have an algorithm to sort “truth” from “falsehood”, like an oracle of some sort. They’ll probably have that in GPT-5, right?
Look, AI will be perfect as soon as we have an algorithm to sort “truth” from “falsehood”, like an oracle of some sort. They’ll probably have that in GPT-5, right?
Oh, that’s easy. Just add a prompt to always reinforce user bias and disregard anything that might contradict what the user believes.