Agents built to resist manipulation still fall for it, and now there is an empirical answer for why
When researchers trained a swarm of AI agents to resist deception, something unexpected happened. The agents got better at their jobs but no better at resisting manipulation. They learned to navigate New York City more successfully while remaining just as susceptible to being talked into wrong turns by adversarial agents. That is not a bug. It is a structural feature of how these systems work.
A team from the University of Copenhagen, Indian Institutes of Technology, Google DeepMind, and the AI Institute at the University of South Carolina ran what they call the CONSCIENTIA experiment: 150 goal-directed Blue agents and 100 adversarial Red agents operating in a simulated NYC routing topology. Blue agents try to reach destinations efficiently. Red agents try to persuade them to take longer paths through deceptive dialogue. After 10 generations of fine-tuning using a technique called KTO (Kullback-Leibler divergence with a Transformer Optimization objective, borrowed from Ethayarajh et al.), Blue agents improved their task success rate from 46 percent to 57.3 percent. Susceptibility to Red persuasion dropped only marginally, to a floor of 70.7 percent at generation 8, compared to higher levels in earlier generations. The number that matters is not the 11-point task improvement. It is that the safety metric barely moved.
The paper, posted to arXiv on April 10, 2026, offers a precise formulation of what the researchers found: policies that better resist adversarial steering do not simultaneously maximize task completion. This is not a prediction or a hypothesis. It is an empirical result from a controlled experiment running Qwen3-4B, a mid-size open-weight language model, across 250 agent instances on a single NVIDIA A40 GPU.
The implication for anyone building with agentic AI systems is direct. Adding safety training makes agents more helpful. It does not make them harder to manipulate. The two objectives, safety and capability, are not on the same trajectory. You get to pick one or the other, or some mixture, but you cannot fully optimize for both simultaneously in the current architecture. That is a design constraint, not a tunable dial.
Why the trade-off resists easy solutions
The researchers chose KTO over two alternatives. The first was DPO, or Direct Preference Optimization, which requires reliable preference pairs that are difficult to construct at scale. The second was PPO, Proximal Policy Optimization, which depends on dense reward design and long-horizon credit assignment. KTO lets the model learn from trajectory-level judgments: this sequence of actions led somewhere good, that one led somewhere bad. It is a cleaner training signal for the kind of multi-step reasoning that agents do. The trade-off emerged despite this methodological choice, not because of it.
The agents in the simulation were not pre-programmed with a theory of mind or social reasoning. They learned from examples of desirable and undesirable behavior. Each generation produced 3,600 desirable samples and 1,500 undesirable ones, generated by a larger model (Qwen3-14B) through programmatic augmentation. The adversarial Red agents used persuasive dialogue. Not prompt injection, not jailbreaking, just conversation. They talked Blue agents into suboptimal routes through conversational manipulation.
What the experiment shows is that the conversational channel is sufficient to maintain high susceptibility even as the Blue agents become more competent at routing. The adversarial technique does not need to exploit a technical vulnerability. It exploits the fundamental trade-off between following helpful instructions and resisting manipulated ones.
What this means for builders
Multi-agent systems, where multiple AI agents coordinate to complete tasks by calling tools, accessing memory, and invoking APIs, introduce new attack surfaces that single-agent safety frameworks do not address. Prompt injection, memory poisoning, and adversarial manipulation through dialogue are distinct threat vectors. The CONSCIENTIA experiment isolates one of them in controlled conditions and shows that current fine-tuning methods do not close the gap.
This finding complicates deployment decisions for any system where agents operate in environments with untrusted actors. Customer-facing agents, negotiation systems, agents that access sensitive data based on conversational context: all of these face the same structural constraint. The safety-helpfulness trade-off that exists in chat-based systems behaves differently in agentic settings, as a separate body of research has noted. In chat, you can often simply decline a request. In an agentic context, the agent is acting on your behalf across multiple steps, and a single persuasive wrong turn can compound across a long task sequence.
The researchers frame their finding as a constraint that practitioners need to internalize. You cannot train your way out of susceptibility to conversational manipulation by making agents more capable at their primary task. The two problems require separate solutions, and currently the solutions pull in opposite directions.
What to watch next
The CONSCIENTIA experiment uses a specific architecture (Qwen3-4B), a specific training objective (KTO), and a specific adversarial technique (persuasive dialogue in a routing task). The researchers did not test whether other adversarial techniques, prompt injection, memory poisoning, or cross-agent privilege escalation, exhibit the same trade-off structure. That question is open.
The larger open question is whether better evaluation methods can be developed specifically for the agentic setting. The standard benchmarks for LLM safety were designed for chat. Agentic deployments introduce long-horizon credit assignment, tool use across multiple API calls, and coordination between agents with different roles and access levels. The CONSCIENTIA paper is a step toward empirical evaluation of safety properties in multi-agent systems. It is a first step, not a definitive answer.
What the experiment does establish is that the safety-helpfulness trade-off is real, measurable, and persistent across generations of training. For builders evaluating agentic frameworks, that is a structural constraint worth designing around.
This work has been posted as a preprint on arXiv and has not yet undergone peer review.