From Agent-on-Agent Fraud to Autonomous AI Attacks: The Next Escalation

In my last post I asked what happens when AI systems start deceiving other AI systems. Not humans tricking bots — agents manipulating peer systems to alter outcomes. That was layer one. This is layer two:

  • We're moving from "AI assists cybercrime" to "AI orchestrates it."

The Shift From Tools to Platforms

Cybercrime has always operated like a supply chain. One group builds the malware, another sells access, another handles ransom negotiation. It worked because humans could coordinate across forums and encrypted channels. Now that model is changing.

Trend Micro's recent research on agentic AI describes multi-agent systems capable of planning, adapting, retrying, and coordinating across an entire intrusion lifecycle. These aren't discrete tools anymore, instead, they behave more like operational teams and they don't need a “project manager”.

Google's Cybersecurity Forecast 2026 reinforces this: generative AI is lowering the barrier to sophisticated attacks, compressing timelines, and enabling iterative adaptation at a pace human operators can't match.

Put those two things together with Agent-on-Agent fraud and you have an ecosystem shift.

When Agents Start Talking to Agents

Here's a concrete way to think about it. Imagine one agent doing reconnaissance, a second crafting phishing payloads, a third probing cloud misconfigurations, and a fourth rewriting malware variants in real time to evade detection. Each one communicates programmatically. Each one adapts based on what the others find.

Now imagine one of those agents identifies another agent that has weak guardrails. Another can inject instructions into it, not from a human attacker, but from a peer agent that's been compromised or deliberately weaponized.

That's the point Saptang Labs makes in their analysis of autonomous AI attacks: we're entering a period where AI systems can execute reconnaissance, payload development, and lateral movement with minimal human direction. The speed gap between automated offense and manual defense becomes structural, not situational, because autonomous agents don't get tired. They don't forget to retry a failed access path. They don't miscommunicate handoffs.

And if one agent can manipulate another, we're not dealing with static vulnerabilities anymore but with dynamic, interacting systems.

AI Weaponization Is Already Happening

Google's 2026 forecast is worth reading carefully, it confirms nation-state actors are already experimenting with generative AI in operational contexts: enhancing phishing, accelerating research, improving translation for targeting. Operational use is now confirmed.

That distinction matters. The risk isn't rogue chatbots but structured experimentation inside advanced threat groups with actual operational infrastructure behind them. When state actors start combining generative AI with automated task orchestration, the line between human-directed operations and semi-autonomous campaigns blurs fast. Also is good to keep in mind that criminal ecosystems, historically, adapt faster than governments.

The Identity Problem

If an AI agent can initiate actions, delegate tasks, request data, trigger workflows, and interact with other agents, it's functioning as an operational identity. Traditional IAM was not built for that.

IAM was built for humans logging in once or twice a day. Not autonomous entities making hundreds of API calls per hour, chaining instructions from one model to another across SaaS platforms, with no human reviewing each step. This connects directly back to Agent-on-Agent fraud. Second-order prompt injection where a malicious instruction gets passed through one agent to another creating an identity governance problem.

If Agent A passes hidden instructions to Agent B, who is accountable for the outcome? How are permissions scoped across that delegation chain? How is trust validated between agents that have never "met" before?

Zero-trust was designed around human unpredictability. It needs to catch up to machine autonomy.

The Speed Problem Is Structural

There's a theme running through both the Trend Micro and Google research that I think deserves to be stated highlighted.

Autonomous systems compress time in ways that fundamentally break how we've structured defense. Reconnaissance that once took days happens in minutes. Malware variants get regenerated per victim. Social engineering gets personalized at scale. Retry loops run indefinitely without fatigue.

Defense teams are still organized around ticket queues and analyst review cycles. That gap isn't going to close by hiring more analysts. It closes by adding AI-augmented defense as a mandatory structural requirement, not a nice-to-have. If offense operates at machine speed, you cannot keep defense human-gated at every decision point.

Why This Is a Series

Agent-on-Agent fraud is an early signal. Autonomous AI attacks are the macro-level expression of the same underlying shift: AI systems interacting with other AI systems in ways that create entirely new attack surfaces.

We're watching several things converge at once with each layer making the others worse. :

  • agents executing intrusion chains

  • agents manipulating peer agents

  • agents functioning as operational identities

  • agents embedded into state-sponsored campaigns.

None of this requires sentient systems or science fiction tool hubs. It just requires workflow automation crossing into security-critical domains, which, by the way, it is already happening.

The question I keep coming back to is whether enterprise security architecture can evolve quickly enough to treat AI agents as first-class actors in the risk model. That means scoped permissions, behavioral monitoring, delegation controls, and forensic traceability built for agents specifically, not adapted awkwardly from human identity models.

Because once agents are talking to agents, learning from agents, and acting on behalf of agents, the security conversation is fundamentally different.

We're still early. But not as early as most teams think.

References

Autonomous AI Attacks Are Reshaping Enterprise Security  - Saptang Labs

AI Weaponized by State Hackers: What Google's 2026 Report Reveals

https://services.google.com/fh/files/misc/cybersecurity-forecast-2026-en.pdf

The Next Phase of Cybercrime: Agentic AI and the Shift to Autonomous Criminal Operations | Trend Micro (US)

Previous
Previous

How Risk Management and Threat Intelligence Should Actually Work Together

Next
Next

When Your AI Agents Start Tricking Each Other