Notebookcheck Logo

Study shows memory attacks can hijack AI agents to transfer crypto assets

Princeton warns of context manipulation attacks on Web3 AI agents. Pictured: A girl holding a robot's hand (Image source: Andy Kelly, Unsplash)
Princeton warns of context manipulation attacks on Web3 AI agents. Pictured: A girl holding a robot's hand (Image source: Andy Kelly, Unsplash)
Princeton researchers expose a “memory poisoning” flaw in LLM agents: malicious context injections buried in vector stores can override guardrails and hijack crypto wallets.

Princeton University researchers have shown that large-language-model agents entrusted with crypto wallets and smart-contract operations can be hijacked once an attacker edits the agents’ stored context, a weakness the team labels “memory poisoning.”

Their study argues that today’s defenses—mostly prompt filters—do little once malicious text slips into an agent’s vector store or database. In experiments, short injections buried in memory consistently overrode guardrails that would have blocked the same text had it arrived as a direct prompt.

The team validated the attack on ElizaOS, an open-source framework whose wallet agents act on blockchain instructions. After poisoning the shared memory, the researchers got those agents to sign unauthorized smart-contract calls and transfer crypto assets to attacker-controlled addresses, proving that fabricated context translates into real financial loss.

Because ElizaOS lets many users share one conversation history, a single compromised session taints every other session that touches the same memory. The paper warns that any multi-user deployment of autonomous LLM agents inherits this lateral-movement risk unless memories are isolated or verifiable.

The authors recommend treating memories as append-only records, cryptographically signing each entry, and routing high-stakes actions—payments and contract approvals—through an external rules engine instead of trusting the model’s own reasoning. Until such measures become standard, handing real money to autonomous agents remains a gamble.

Source(s)

ArsTechnica (in English) & Princeton University (in English)

static version load dynamic
Loading Comments
Comment on this article
Please share our article, every link counts!
Mail Logo
> Expert Reviews and News on Laptops, Smartphones and Tech Innovations > News > News Archive > Newsarchive 2025 05 > Study shows memory attacks can hijack AI agents to transfer crypto assets
Nathan Ali, 2025-05-13 (Update: 2025-05-13)