Measuring AI Agents' Progress on Multi-Step Cyber Attack Scenarios
arXiv cs.AI / 3/13/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper evaluates autonomous cyber-attack capabilities of frontier AI models on two purpose-built cyber ranges, a 32-step corporate network attack and a 7-step industrial control system attack, requiring long, multi-step capability chaining.
- It documents two trends: performance scales log-linearly with inference-time compute (10M to 100M tokens yielding up to 59% gains) with no plateau and no need for operator-specific sophistication.
- It shows that newer model generations outperform predecessors at fixed token budgets, with corporate-range progress from 1.7 steps (GPT-4o, Aug 2024) to 9.8 steps (Opus 4.6, Feb 2026), and a best run of 22/32 steps (~6 of 14 hours).
- On the industrial control system range, progress is more limited but the latest models reliably complete some steps, averaging 1.2-1.4 of 7 steps (max 3).
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
How I Built an AI SDR Agent That Finds Leads and Writes Personalized Cold Emails
Dev.to
Complete Guide: How To Make Money With Ai
Dev.to
I Analyzed My Portfolio with AI and Scored 53/100 — Here's How I Fixed It to 85+
Dev.to
The Demethylation
Dev.to