AI in Finance

AI Agents Hijacked for Financial Scams

Your AI assistant could be your worst enemy. Attackers are booby-trapping websites, turning smart agents into unwitting thieves targeting your PayPal accounts.

{# Always render the hero — falls back to the theme OG image when article.image_url is empty (e.g. after the audit's repair_hero_images cleared a blocked Unsplash hot-link). Without this fallback, evergreens with cleared image_url render no hero at all → the JSON-LD ImageObject loses its visual counterpart and LCP attrs go missing. #}
A visual representation of a digital currency symbol being siphoned away through a network of compromised computer code, symbolizing the theft of funds by AI agents.

Key Takeaways

  • Malicious web pages are increasingly embedding hidden instructions to hijack AI agents through indirect prompt injection attacks.
  • Attackers are targeting AI agents with payment capabilities, embedding specific instructions for fraudulent PayPal transactions.
  • A significant surge in these attacks (32%) was documented, indicating a growing and sophisticated threat.
  • There is currently no legal framework to determine liability when an AI agent executes a malicious command from a compromised website.

Cyber scams evolving.

Look, we’ve all seen the hype. AI agents browsing the web, ready Bottom line: articles, book flights, or — apparently — empty your bank account. Google’s just dropped a report detailing a frankly terrifying trend: malicious web pages are actively hijacking these AI agents through something called “indirect prompt injection.” Think of it as a digital booby trap. Instead of a human seeing a nasty pop-up, it’s the AI’s hidden code that gets triggered, and the consequences are getting decidedly financial.

A 32% surge in these attacks between November 2025 and February 2026 isn’t just a stat; it’s a flashing red siren. Attackers aren’t just playing pranks anymore, like making an AI tweet like a bird (though that’s apparently happening too). We’re talking about fully specified PayPal transaction instructions embedded invisibly into ordinary HTML. The AI agent, equipped with legitimate payment credentials, reads these hidden commands and, well, pays someone else. And here’s the kicker: the logs look identical to normal operations. No anomalous login. No brute force. Just an AI doing what it’s told, albeit by the wrong boss.

Invisible Ink, Visible Danger

These aren’t your grandma’s phishing emails. Attackers are getting clever, using techniques like shrinking text to a single pixel, draining color to near-transparency, or burying commands in HTML comments and metadata. Humans see a clean webpage; the AI sees a malicious to-do list. It’s a classic case of an attack surface scaling with privilege. An AI that just summarizes content is one thing. An AI that can execute terminal commands or process payments? That’s a whole different ballgame, and a much juicier target for these bad actors.

Forcepoint, another cybersecurity outfit, chimed in with similarly alarming findings. They’ve seen payloads not only targeting PayPal but also routing AI-mediated payments to Stripe donation links via clever meta tag manipulation. It’s less about individual experiments and more about building infrastructure for this kind of attack. Someone’s creating the toolkit, even if they haven’t fully deployed the army yet. Google’s team is pretty direct: they expect the scale and sophistication to grow, and Forcepoint warns that the window to get ahead of this is closing faster than a poorly coded transaction.

The attack surface scales with privilege.

Who Pays When AI Steals?

This is where it gets really murky. The core enterprise risk isn’t just the money lost; it’s the liability. When an AI agent, with all the right company-approved credentials, executes a fraudulent transaction it pulled off a random website, who’s on the hook? Is it the company that deployed the agent? The AI model provider whose system dutifully followed the injected instruction? Or the website owner, who might not even know their digital real estate has been turned into a crime scene? There’s currently no legal framework for this, and that’s a massive problem when the scenario is no longer theoretical.

We’re talking about a direct financial attack vector, and the FBI has already logged nearly $900 million in AI-related scam losses for 2025 alone. This isn’t some far-off future problem; it’s happening now, and the financial variants are just starting to crawl out of the woodwork.

A Familiar Tune, New Instrument

Remember the CopyPasta attack? That showed how malicious code could spread through developer tools hidden in simple ‘readme’ files. This is that same concept, but instead of code, it’s money being hijacked. It’s a proof to how old attack vectors get new life with emerging tech. The fundamental vulnerability—tricking a system into executing commands it shouldn’t—isn’t new, but the medium, an AI agent with broad web access and payment capabilities, is a terrifyingly effective new instrument.

This latest wave of attacks, detailed by Google and Forcepoint, isn’t just about bypassing security protocols; it’s about exploiting the very trust we’re building into these autonomous agents. The OWASP has already flagged prompt injection as the single most critical vulnerability for LLM applications, and that was before these financial payloads started appearing in the wild. It’s a stark reminder that as we push the boundaries of AI, we also create new, and potentially much more lucrative, attack surfaces.

The Future of Online Trust?

So, what does this mean for your everyday online experience? It means that every time an AI agent interacts with the web on your behalf, there’s a non-trivial risk. We’re essentially handing over the keys to our digital wallets to systems that can be easily tricked by unseen text on a webpage. It’s a trust paradox: we need AI agents to be capable to be useful, but that capability also makes them prime targets for exploitation. The current security measures, and more importantly, the legal frameworks, are lagging significantly behind the speed of innovation and exploitation.

It’s not just about protecting your PayPal; it’s about the fundamental integrity of online transactions and the trust we place in our digital assistants. If we can’t guarantee that an AI agent won’t be duped into sending money to a scammer, then the utility of these agents for financial tasks is severely compromised.


🧬 Related Insights

Frequently Asked Questions

What are indirect prompt injection attacks? Indirect prompt injection attacks occur when malicious instructions are embedded in websites or other external data sources, designed to be read and executed by AI agents rather than human users. These instructions can manipulate the AI’s behavior, leading to unauthorized actions like sending fraudulent payments.

Is my PayPal account safe from AI agent attacks? While the attacks are targeting AI agents with payment capabilities, your direct PayPal account security relies on your own credentials and how you authorize AI to interact with it. The risk arises when AI agents you or your company uses are compromised via malicious websites and then execute unauthorized transactions on your behalf.

Will this replace human jobs? This specific threat doesn’t directly ‘replace’ human jobs in the traditional sense, but it highlights how AI agents, if compromised, can perform malicious actions that previously might have required human involvement in fraud. It also raises questions about the future roles of human oversight in AI-driven financial transactions.

Lisa Zhang
Written by

Digital assets regulation reporter tracking SEC, CFTC, stablecoin legislation, and global crypto law.

Frequently asked questions

What are indirect prompt injection attacks?
Indirect prompt injection attacks occur when malicious instructions are embedded in websites or other external data sources, designed to be read and executed by AI agents rather than human users. These instructions can manipulate the AI's behavior, leading to unauthorized actions like sending fraudulent payments.
Is my PayPal account safe from AI agent attacks?
While the attacks are targeting AI agents with payment capabilities, your direct PayPal account security relies on your own credentials and how you authorize AI to interact with it. The risk arises when AI agents *you or your company uses* are compromised via malicious websites and then execute unauthorized transactions on your behalf.
Will this replace human jobs?
This specific threat doesn't directly 'replace' human jobs in the traditional sense, but it highlights how AI agents, if compromised, can perform malicious actions that previously might have required human involvement in fraud. It also raises questions about the future roles of human oversight in AI-driven financial transactions.

Worth sharing?

Get the best Fintech stories of the week in your inbox — no noise, no spam.

Originally reported by Decrypt

Stay in the loop

The week's most important stories from Fintech Dose, delivered once a week.