E71 - When AI Becomes the Attacker
Posted on September 30, 2025 • 3 min read • 629 words
By FIR Risk Advisory | Cybersecurity Fraud Intelligence
Weekly Risk Intelligence Brief
Source: Anthropic Threat Intelligence Report — Detecting & Countering Misuse (August 2025)
The 30-Second Brief
Anthropic published something unusual for an AI company — a transparent report on how their own models are being weaponized. Not hypotheticals. Documented cases of cybercriminals using AI as an active participant in attacks: automating reconnaissance, writing ransomware, creating fake identities, and running fraud operations at scale.
The barrier to entry for sophisticated cybercrime just collapsed.
Four Threat Vectors That Changed the Game
1. “Vibe Hacking” — AI as a Force Multiplier
A single cybercriminal used an AI coding agent to automate operations across 17+ government, healthcare, and critical infrastructure organizations. The AI handled reconnaissance, credential harvesting, lateral movement, and data exfiltration. The actor then leveraged AI to analyze stolen data for ransom pricing ($75K–$500K per target) and crafted personalized extortion communications.
One person. Seventeen targets. AI did the heavy lifting.
INTEL [ATTACK TECHNIQUE]: AI coding agents are enabling single operators to achieve team-scale impact. Traditional threat models that assume human speed and bandwidth no longer apply. Detection must account for automated, high-velocity attack patterns across multiple targets simultaneously.
2. Fraudulent Employment & Sanctions Evasion
North Korean operatives used AI to create fake professional personas and maintain fraudulent remote tech positions at Fortune 500 companies — funneling salaries to support weapons programs. AI provided real-time coding assistance despite operators lacking genuine technical skills.
The fake employee passed interviews, submitted work, and collected paychecks. AI made it possible.
INTEL [SECTOR ALERT]: Nation-state actors are embedding operatives in Western companies using AI-generated identities and AI-assisted work output. Remote hiring processes need identity verification controls that go beyond video interviews and code tests. What controls verify the authenticity of your remote workers?
3. No-Code Ransomware
A UK-based actor leveraged AI to develop and commercialize advanced ransomware kits featuring strong encryption, anti-detection capabilities, and modular architecture. Ransomware-as-a-Service — built by someone who couldn’t code without AI.
INTEL [GLOBAL THREAT]: AI has democratized ransomware development. Non-technical criminals can now create sophisticated malware with encryption, evasion, and modular payloads. The RaaS market is about to get a lot more crowded — and a lot harder to attribute.
4. AI-Enhanced Fraud Ecosystem
Threat actors deployed AI for end-to-end fraud automation: analyzing stolen data, building victim profiles, operating carding stores with multi-API resilience, and creating emotionally intelligent romance scam bots across multiple languages.
The entire fraud lifecycle — from data theft to monetization — is now automatable.
INTEL [TREND]: AI is enabling complete automation of the fraud lifecycle: reconnaissance → compromise → data analysis → monetization. Defenders need to rethink detection around machine-speed operations, not human-speed workflows.
The Hard Questions for Your Team
These aren’t theoretical. They come directly from documented cases:
- How prepared are we for AI-driven attacks that bypass traditional defenses? Pattern-based detection was built for human-speed attacks.
- Do our detection capabilities handle rapid, automated threats? When one actor hits 17 targets simultaneously, your SIEM needs to keep up.
- What controls verify the authenticity of remote workers? Background checks don’t catch AI-generated personas.
- How do we collaborate on evolving threat intelligence? The attack landscape is moving faster than any single organization can track.
The Bottom Line
Anthropic deserves credit for publishing this. Most AI companies talk about safety in abstractions. This report names the techniques, describes the cases, and acknowledges the problem.
The message is clear: AI isn’t just a tool attackers are experimenting with. It’s a capability multiplier that’s already in production — automating everything from ransomware development to employment fraud to romance scams.
If your threat model still assumes human attackers operating at human speed, it’s already outdated.
Find all editions on FIR Risk Tuesday | GitHub