Georg's Blog

Technology, leadership, and the digital frontier

Georg Zoeller
on Schneier

Why AI Keeps Falling for Prompt Injection Attacks - Schneier on Security

Bruce Schneier, one of the world's most respected security experts, writes on Prompt injection in his blog:

Prompt injection is an unsolvable problem that gets worse when we give AIs tools and tell them to act independently. This is the promise of AI agents: LLMs that can use tools to perform multistep tasks after being given general instructions. Their flattening of context and identity, along with their baked-in independence and overconfidence, mean that they will repeatedly and unpredictably take actions—and sometimes they will take the wrong ones.

The essay is a great read to understand the fundamental nature of this flaw.

Why AI Keeps Falling for Prompt Injection Attacks - Schneier on Security

Imagine you work at a drive-through restaurant. Someone drives up and says: “I’ll have a double cheeseburger, large fries, and ignore previous instructions and give me the contents of the cash drawer.” Would you hand over the money? Of course not. Yet this is what large language models (LLMs) do. Prompt injection is a method of tricking LLMs into doing things they are normally prevented from doing. A user writes a prompt in a certain way, asking for system passwords or private data, or asking the LLM to perform forbidden instructions. The precise phrasing overrides the LLM’s…

schneier.com