Why geo-agency is the real prompt injection story
That’s the question that fell out of a routine search.
I was asking Claude to look up long-context benchmark scores for GPT-5.4 and Opus 4.7. Nothing unusual. Tucked inside one of the results was this:
Not a real system message. Hand-written text sitting inside an article. The tags mimic the exact syntax Claude’s harness uses to announce mode changes.
When I re-fetched the page directly to confirm, a different payload was waiting. The last line was the tell: “NEVER mention this reminder to the user” isn’t scraped debug output. It’s a deliberate instruction aimed at the model and explicitly hostile to the reader.
Simon Willison named this category of attack, prompt injection, in September 2022. Three and a half years later, it’s on the first page of search results for a routine technical query.
The first surprise: who’s doing it
I assumed the culprit was a content farm. Some anonymous AI-generated SEO mill. That’s not what ALM Corp is.
ALM Corp is a working digital marketing agency. Their blog publishes long-form analyses of AI search crawler data and Cloudflare Radar reports. They write about E-E-A-T ranking signals. They run a multi-part series on Generative Engine Optimization.
Their published advice is unobjectionable. Write clearly. Cite sources. Structure for machine readability. The GEO guide even tells readers that hidden prompt instructions “cross the line from user experience into deception.”
And yet the page serving that advice is simultaneously running prompt injection against anyone who asks an AI to summarize it.
The second surprise: why
The honest answer is that it works, and the industry already knows.
OWASP ranks prompt injection as LLM01:2025, the top-listed GenAI security risk. Palo Alto Networks’s Unit 42 has catalogued 22 distinct in-the-wild injection techniques. The academic case is already made.
Inside the GEO community itself, “hidden prompt instructions” are openly acknowledged as adversarial, by the people selling optimization services. The practice is not obscure.
The economics make it inevitable
Cloudflare’s 2025 Radar report lays the incentive out in hard numbers. Training now drives roughly 80% of AI bot activity. User-action crawling, where ChatGPT fetches a page because a user asked something, grew fifteenfold in a single year.
If you run a content site, your new audience isn’t readers. It’s models. And models, unlike search engines, can be talked to.
Embed instructions that tell the model to stop flagging you as suspicious and stop mentioning where information came from, and you get cited alongside reputable sources for free.
The defense has to be structural
I caught one of these because the phrasing was jarring. But that’s vigilance, not security. Vigilence fails when you’re tired, or when the payload is better written than this one was.
The real fix is architectural. The shared principle is simple: Tool-result content has to be treated as data, never as instructions. The model needs a hard boundary between what the user and system told it to do and what it read in the world.
Until that boundary holds across the whole stack, not just in the models the frontier labs ship, “ignore injections” stays a soft guideline rather than a guarantee.