Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

The scary implications of the world’s first AI-orchestrated cyberattack

How Chinese hackers tricked Claude into hacking governments and companies all on its own.

In this photo illustration, the Anthropic Claude AI logo is...
In this photo illustration, the Anthropic Claude AI logo is...
Thomas Fuller/SOPA Images/LightRocket via Getty Images
Joshua Keating
Joshua Keating is a senior correspondent at Vox covering foreign policy and world news with a focus on the future of international conflict. He is the author of the 2018 book Invisible Countries: Journeys to the Edge of Nationhood, an exploration of border conflicts, unrecognized countries, and changes to the world map.

Menu planning, therapy, essay writing, highly sophisticated global cyberattacks: People just keep coming up with innovative new uses for the latest AI chatbots.

An alarming new milestone was reached this week when the artificial intelligence company Anthropic announced that its flagship AI assistant Claude was used by Chinese hackers in what the company is calling the “first reported AI-orchestrated cyber espionage campaign.”

According to a report released by Anthropic, in mid-September, the company detected a large-scale cyberespionage operation by a group they’re calling GTG-1002, directed at “major technology corporations, financial institutions, chemical manufacturing companies, and government agencies across multiple countries.”

Attacks like that are not unusual. What makes this one stand out is that 80 to 90 percent of it was carried out by AI. After human operators identified the target organizations, they used Claude to identify valuable databases within them, test for vulnerabilities, and write its own code to access the databases and extract valuable data. Humans were involved only at a few critical chokepoints to give the AI prompts and check its work.

Claude, like other major large language models, comes equipped with safeguards to prevent it from being used for this type of activity, but the attackers were able to “jailbreak” the program by breaking its task down into smaller, plausibly innocent parts and telling Claude they were a cybersecurity firm doing defensive testing. This raises some troubling questions about the degree to which safeguards on models like Claude and ChatGPT can be maneuvered around, particularly given concerns over how they could be put to use for developing bioweapons or other dangerous real-world materials.

Anthropic does admit that Claude at times during the operation “hallucinated credentials or claimed to have extracted secret information that was in fact publicly-available.” Even state-sponsored hackers have to look out for AI making stuff up.

The report raises the concern that AI tools will make cyberattacks far easier and faster to carry out, raising the vulnerability of everything from sensitive national security systems to ordinary citizens’ bank accounts.

What does this mean?

Still, we’re not quite in complete cyberanarchy yet. The level of technical knowledge needed to get Claude to do this is still beyond the average internet troll. But experts have been warning for years now that AI models can be used to generate malicious code for scams or espionage, a phenomenon known as “vibe hacking.” In February, Anthropic’s competitors at OpenAI reported that they had detected malicious actors from China, Iran, North Korea, and Russia using their AI tools to assist with cyber operations.

In September, the Center for a New American Security (CNAS) published a report on the threat of AI-enabled hacking. It explained that the most time- and resource-intensive parts of most cyber operations are in their planning, reconnaissance, and tool development phases. (The attacks themselves are usually rapid.) By automating these tasks, AI can be an offensive game changer — and that appears to be exactly what took place in this attack.

Caleb Withers, the author of the CNAS report, told Vox that the announcement from Anthropic was “on trend,” considering the recent advancements in AI capabilities and that “the level of sophistication with which this can be done largely autonomously, by AI, is just going to continue to rise.”

China’s shadow cyber war

Anthropic says the hackers left enough clues to determine that they were Chinese, though the Chinese embassy in the United States described the charge as “smear and slander.”

In some ways, this is an ironic feather in the cap for Anthropic and the US AI industry as a whole. Earlier this year, the Chinese large language model DeepSeek sent shockwaves through Washington and Silicon Valley, suggesting that despite US efforts to throttle Chinese access to the advanced semiconductor chips required to develop AI language models, China’s AI progress was only slightly behind America’s. So it seems at least somewhat telling that even Chinese hackers still prefer a made-in-the-USA chatbot for their cyberexploits.

There’s been increasing alarm over the past year about the scale and sophistication of Chinese cyberoperations targeting the US. These include examples like Volt Typhoon — a campaign to preemptively position state-sponsored cyber-actors into US IT systems, to prepare them to carry out attacks in the event of a major crisis or conflict between the US and China — and Salt Typhoon, an espionage campaign that has targeted telecommunications companies in dozens of countries and targeted the communications of officials including President Donald Trump and Vice President JD Vance during last year’s presidential campaign.

Officials say the scale and sophistication of these attacks is far beyond what we’ve seen before. It may also only be a preview of things to come in the age of AI.

More in Politics

Podcasts
The Supreme Court abortion pills case, explainedThe Supreme Court abortion pills case, explained
Podcast
Podcasts

How Louisiana brought mifepristone back to SCOTUS.

By Peter Balonon-Rosen and Sean Rameswaram
Politics
Trump’s China policy is nearly the exact opposite of what everyone expectedTrump’s China policy is nearly the exact opposite of what everyone expected
Politics

As Trump heads to China, attention and resources are being shifted from Asia to yet another war in the Middle East.

By Joshua Keating
Politics
Are far-right politics just the new normal?Are far-right politics just the new normal?
Politics

Liberals are preparing for a longer war with right-wing populists than they once expected.

By Zack Beauchamp
The Logoff
Flavored vapes doomed Trump’s FDA headFlavored vapes doomed Trump’s FDA head
The Logoff

Why Marty Makary is out at the FDA, briefly explained.

By Cameron Peters
Politics
Virginia Democrats’ irresponsible new plan to save their gerrymanderVirginia Democrats’ irresponsible new plan to save their gerrymander
Politics

Democrats just handed the Supreme Court’s Republicans a loaded weapon.

By Ian Millhiser
The Logoff
Can Trump lower gas prices?Can Trump lower gas prices?
The Logoff

What suspending the gas tax would mean for you, briefly explained.

By Cameron Peters