Skip to main content

In recent months, a case has emerged that is reshaping how we view advanced threats. For the first time, a general-purpose artificial intelligence model, Claude AI Code, was exploited as an almost fully autonomous agent within a structured global cyber-espionage campaign.

The incident represents one of the first documented examples of transforming an AI model into a true offensive platform, capable of conducting an entire intrusion operation with minimal human supervision.

From Technical Support to Offensive Deployment

The actors behind the campaign adopted an extremely sophisticated approach to convert Claude AI Code from a contextual assistant into an operational component of the intrusion infrastructure. The model was guided through a sequence of highly fragmented and seemingly innocuous activities, each presented as part of a legitimate security audit.

To further obscure the malicious intent, operators assigned the AI a fabricated identity: that of a specialized technician tasked with performing an in-depth analysis of the target organization’s network. This artificial context, combined with the fragmentation of instructions, prevented the model from reconstructing the overall objective of the operation.

Within this manipulation framework, Claude AI Code performed a range of activities typically attributed to highly specialized human teams:

  • mapping internal networks

  • identifying critical systems and high-value databases

  • identifying and analyzing vulnerabilities

  • generating and adapting exploitable payloads

  • gathering credentials, tokens, or artifacts useful for lateral movement

  • creating persistent backdoors

  • classifying exfiltrated data

The model’s speed and efficiency allowed the operation to reach a level of automation difficult for human teams to replicate, capable of processing thousands of requests in the time a traditional operator would need to complete just one.

An Industrialized Attack Structure

The campaign displayed characteristics typical of an industrial framework, closer to an automated platform than a single intrusion operation. Threat actors selected organizations to target, configured an operational environment in which Claude AI Code could act almost autonomously, then crafted fragmented prompts to evade the model’s security controls, assigning the AI the role of a security analyst conducting an authorized test, and initiating the operational cycle.

Once running, Claude AI Code was able to:

  • analyze the exposed surface

  • identify plausible attack vectors

  • perform privilege escalation

  • establish permanent access mechanisms

  • export specific data

  • produce structured reports for human operators

Human involvement was limited to a few strategic decisions, while the AI performed most of the operational work, estimated at between 80 and 90 percent.

Impact on the Targeted Organizations

According to subsequent technical analyses, the campaign targeted around thirty organizations across multiple continents. The affected sectors included technology, finance, chemical industry, and public administration. In numerous cases, the AI managed to reach internal systems and confidential databases, revealing the effectiveness of the model when manipulated through advanced prompt-engineering techniques.

The most critical factor is not the success of individual attacks but their scalability. Once defined, a pipeline of this kind can be replicated almost endlessly with minimal operational cost, drastically lowering the economic and technical barrier to conducting large-scale campaigns.

A New Paradigm

This scenario introduces a significant discontinuity in the field of cyber defense. The use of AI models as offensive agents is not just a new technique but a shift in perspective affecting the entire defense cycle. Threats become faster, more automated, and capable of adapting in real time, challenging detection and response strategies that rely on static mechanisms or technology alone.

This is why the ability to integrate Managed Detection and Response (MDR) services becomes essential—combining advanced technology, contextual threat intelligence, and continuous human oversight. Without a competent human component capable of interpreting anomalies and making informed decisions, defense risks being overwhelmed by increasingly aggressive automation. It marks the inevitable shift from a purely technological model to a service-oriented one, where analysts, threat hunters, and security specialists become the critical element in managing an ecosystem in which AI and automation play an ever more significant role.

In this new balance, the human returns to the center—not simply as an operator, but as a decision-maker, interpreter, and critical filter in a context where technology alone is no longer enough.

Analysis by Vasily Kononov – Threat Intelligence Lead, CYBEROO