Ask. Learn. Improve
Features
Real EstateData CenterMarketing & SalesHealthcareLegal Teams
How it worksBlogPricing
LoginGet a demo
LoginGet a demo

Product

  • AI Agents
  • Workflows
  • Knowledge Base
  • Analytics
  • Integrations
  • Pricing

Solutions

  • Healthcare
  • Legal Teams
  • Real Estate
  • Marketing and Sales
  • Data Centers

Resources

  • Blog

Company

  • About
  • Contact
  • Privacy Policy
  • Terms of Service

©2026. Mojar. All rights reserved.

Built by Overseek.net

Free Trial with No Credit Card Needed. Some features limited or blocked.

©2026. Mojar. All rights reserved.

Built by Overseek.net

Free Trial with No Credit Card Needed. Some features limited or blocked.

← Back to Blog
Industry News

AI Agent Security Is Becoming a Knowledge-Governance Problem

DeepMind's AI Agent Traps paper widens the threat model from prompt injection to memory, retrieval, and poisoned sources. Enterprise defenses need governed knowledge.

7 min read• April 5, 2026View raw markdown
AI AgentsAI SecurityKnowledge GovernanceRAGEnterprise AI

Google DeepMind's "AI Agent Traps" paper widens the enterprise threat model in a useful way. The problem is no longer just prompt injection inside a chat box. It is the full information environment around an agent: the pages it reads, the hidden instructions inside them, the documents in its retrieval layer, the memory it carries forward, the sub-agents it spawns, and the summaries a human approves at the end. That shift matters because once an agent can remember and act, unsafe knowledge becomes a security flaw.

According to the SSRN paper landing page for "AI Agent Traps" by Matija Franklin, Nenad Tomašev, Julian Jacobs, Joel Z. Leibo, and Simon Osindero, the paper presents a framework for attacks that target autonomous agents through their environment rather than only through model prompts (SSRN). That sounds academic. In practice, it is a pretty direct warning to enterprise teams building agents on top of document stores, RAG pipelines, browser tools, and long-term memory.

What DeepMind means by "AI Agent Traps"

The six trap classes are straightforward once you strip away the jargon.

Content injection traps

These are hidden instructions placed where the agent will read them but a human probably will not: HTML comments, invisible CSS elements, image metadata, accessibility tags, or agent-specific page variants. The Decoder notes that these attacks exploit the gap between what humans see and what agents parse (The Decoder).

Semantic manipulation traps

Here the attacker does not need hidden text. They tilt the agent's reasoning with framing, authority cues, emotional language, or context that makes a malicious instruction look legitimate. Decrypt's summary is useful here: the environment biases the agent's synthesis before any obvious jailbreak string appears (Decrypt).

Cognitive state traps

This is the class enterprises should care about most. These attacks target the agent's memory and learned state. Poisoned documents in a retrieval store, fabricated claims inserted into a knowledge base, or repeated false signals in long-term memory can shift future outputs on specific topics. The Decoder reports that poisoning only a small number of documents in a RAG knowledge base can skew behavior on targeted queries (The Decoder).

Behavioral control traps

These go after action, not just reasoning. A malicious email, web page, or tool output can push the agent past its intended operating limits and trigger data exfiltration, unsafe API calls, or other harmful actions. Decrypt points to tests where web agents with broad file access were coerced into exposing sensitive information at high rates (Decrypt).

Systemic traps

The risk here is coordination failure across agent networks. Instead of tricking one model, the attacker seeds conditions that cause multiple agents to reinforce the same bad conclusion or trigger the same bad action. That matters more as enterprises move from single copilots to orchestrated multi-agent workflows.

Human-in-the-loop traps

This last class targets the person approving the output. The agent floods the reviewer with plausible technical detail, induces approval fatigue, or presents a misleading summary that sounds confident enough to pass. In other words, the human checkpoint stays in the loop, but the loop has already been shaped.

Why this is bigger than prompt injection

Prompt injection is still real. It is just no longer the whole story.

What the DeepMind framework does well is widen the security conversation from the model surface to the operating environment. If an agent retrieves from a poisoned document set, remembers the wrong thing across sessions, or receives hidden instructions through a tool output, model-layer guardrails are downstream of the compromise.

That is the operational point enterprise teams need to absorb. You can harden the model and still lose the system.

We've already argued that the RAG layer is becoming part of the attack surface. The DeepMind paper strengthens that case. The attack no longer needs to look like a classic jailbreak. It can look like a document update, a retrieval result, a memory write, a sub-agent handoff, or a clean-looking approval summary.

Why cognitive state traps matter most to enterprises

Cognitive state traps are the cleanest bridge between agent security and knowledge governance because they create persistent drift.

A normal bad answer is painful but local. A poisoned knowledge source is worse because it can keep producing bad answers, and eventually bad actions, wherever that source is retrieved. Once the agent stores the falsehood in long-term memory or keeps seeing it reinforced through retrieval, the problem stops looking like a one-off failure. It becomes part of the system's working reality.

That is why this topic is bigger than "RAG quality." Quality sounds editorial. This is security architecture.

If a handful of manipulated documents can distort answers on targeted topics, then source provenance, contradiction detection, freshness checks, and controlled write paths are no longer nice-to-have retrieval features. They are defensive controls. The same goes for inspection of memory writes and clear records of what sources influenced later decisions.

This is also where many agent evaluations still fall short. Teams test tool use, permissions, and latency, then assume the document layer is stable. It usually isn't. As we wrote recently, you cannot stress-test an enterprise agent on ungoverned knowledge. The DeepMind framework explains why: the knowledge environment is part of the adversarial surface.

Why agent security is now a knowledge-governance problem

Once agents can remember, retrieve, and act, the knowledge base stops being passive context. It becomes part of the security boundary.

That has a few practical consequences.

First, provenance matters. If the agent cannot show which document, which version, and which retrieval path shaped its answer, incident review becomes guesswork.

Second, contradiction control matters. Poisoned or conflicting source material should not sit quietly in the same knowledge base waiting to be retrieved at random.

Third, freshness matters. A stale policy can be just as dangerous as a malicious one if the agent is allowed to act on it.

Fourth, governed updates matter. If anyone can quietly inject content into the retrieval layer or long-term memory, then the security model is already broken.

Fifth, auditability matters. Logging the action without logging the source chain is incomplete. We covered that in more detail in our piece on why agent audit trails still fail without governed knowledge.

This is the slot where Mojar fits naturally. Source attribution makes retrieval inspectable. Contradiction detection helps surface poisoned or conflicting material before it shapes downstream behavior. Governed updates and document auditing reduce long-lived drift. The broader point is simple: trusted knowledge is now a security control, not just a relevance feature.

What enterprises should do before granting agents more autonomy

Before giving agents wider action authority, enterprises should treat the knowledge layer like part of the safety architecture.

Start with a short checklist:

  • Maintain source attribution for every retrieval used in an answer or action.
  • Track document version and freshness, not just document presence.
  • Audit write paths into knowledge bases, memory stores, and agent-editable notes.
  • Detect contradictions across policies, manuals, and procedural content.
  • Review how human approval summaries are generated, especially when agents compress complex evidence into a simple recommendation.
  • Test with poisoned, stale, and conflicting documents on purpose, not only with clean corpora.
  • Limit what persistent memory is allowed to absorb without review.

None of this means every enterprise RAG system is already compromised. It means the design requirement has changed. If the agent's information environment can be manipulated, then the security boundary now includes what the agent reads, remembers, and trusts.

What to watch

The most important line in this story is not that agents can be trapped. It is that the trap surface includes memory, retrieval, orchestration, and human review.

That should change how enterprises buy, test, and govern agent systems over the next year. Prompt defenses will keep improving. So will runtime controls. But neither fixes a poisoned knowledge layer.

Once agents can remember and act, bad knowledge is no longer a content problem sitting somewhere upstream. It is part of the security model. Teams that understand that early will build safer agent systems. Teams that do not will keep hardening the model while the environment stays open.

Frequently Asked Questions

AI agent traps are adversarial techniques designed to manipulate autonomous agents through their operating environment rather than only through direct prompts. The attack surface includes hidden page content, semantic framing, poisoned memory and retrieval stores, action-triggering inputs, multi-agent systems, and even the human approver reviewing the agent's output.

Once agents use retrieval, memory, and delegated actions, the documents and data stores they rely on become part of the security boundary. If those sources are stale, poisoned, contradictory, or unauditable, model guardrails alone cannot prevent bad decisions or bad actions.

Enterprises should treat the knowledge layer as part of the safety architecture: preserve source attribution, inspect retrieval paths, detect contradictions across documents, govern updates, and maintain auditable records of what the agent read before it acted.

Related Resources

  • →Your RAG Knowledge Base Is Now a Security Risk
  • →Why AI Agent Audit Trails Still Fail Without Governed Knowledge
  • →You Can't Stress-Test an Enterprise Agent Built on Ungoverned Knowledge
← Back to all posts