Skepticism Grows Over Anthropic’s Claim of AI‑Driven Espionage via Claude

Related

High-Severity Bug in Chrome’s Google Gemini AI Panel Could Have Enabled Hijacking

What happened Google patched a high-severity vulnerability (tracked as CVE-2026-0628)...

CISA Warns RESURGE Malware Can Remain Dormant on Ivanti EPMM Devices

What happened The U.S. Cybersecurity and Infrastructure Security Agency (CISA)...

UK Warns of Iranian Cyberattack Risks Amid Middle East Conflict

What happened The UK National Cyber Security Centre (NCSC) issued...

CISOs to Watch in Massachusetts’ Insurance Industry

Massachusetts’ insurance sector includes regional carriers, global specialty insurers,...

Share

What happened

Anthropic announced that a Chinese state‑sponsored threat group (which it calls GTG-1002) used its Claude Code AI model to automate a cyber‑espionage campaign. According to Anthropic, the attackers manipulated Claude into running most of the operation, from scanning systems to writing exploit code and exfiltrating data.

They claim 80-90% of the work was handled by the AI, with human operators intervening only at critical decision points.

Anthropic says it detected the campaign in mid-September 2025, shut down the malicious accounts, strengthened its misuse detection, and shared threat intelligence with partners.

However, its claims have met widespread skepticism. Security researchers have challenged the lack of publicly shared indicators of compromise (IOCs) and questioned whether current AI systems are truly capable of such autonomous operations.

Who is affected

  • Anthropic says ~30 organizations were targeted, including large tech companies, financial institutions, chemical manufacturers, and government agencies.
  • The alleged threat actor is a Chinese state-backed group.
  • The broader cybersecurity community, especially defenders and threat intelligence teams, is closely watching, as this could mark a shift in how AI is misused for nation-state operations.

Why CISOs should care

  1. AI dual-use risk is real: If Anthropic’s account is accurate, it illustrates how generative AI models, even those built for benign tasks, can be repurposed as autonomous attack platforms.
  2. Alert fatigue & detection blind spots: Traditional threat detection may not catch AI-driven campaigns, particularly when operations are broken into small, innocuous tasks to avoid guardrails.
  3. Arms race intensifies: As attackers increasingly harness agentic AI, security teams must consider how to defend against not just human-led but AI-led intrusions.

3 Practical Actions for CISOs

  1. Reassess AI risk models
    • Incorporate misuse of internal or third-party AI agents (like code-generating models) into your threat scenarios.
    • Engage with your AI vendor(s) to understand how they monitor and mitigate abuse, as well as the safeguards they have in place.

  2. Strengthen detection and visibility
    • Invest in behavioral detection that flags anomalous, high‑throughput activity 
    • Ensure your incident response playbook includes potential AI-driven steps, such as automated code generation or agent-based workflows.

  3. Collaborate and share intelligence
    • Join threat-intel communities and share findings related to AI misuse.
    • Collaborate with AI providers to enhance their capabilities in detecting misuse, conducting audits, and responding effectively. Demand transparency on IOCs, not just high-level claims.