GPT-5.5 Shows Autonomous Cyberattack Skills, Matching Claude Mythos
UK's AI Security Institute finds GPT-5.5 can autonomously perform complex cyberattacks, solving a 12-hour human puzzle in 10 minutes. The model's offensive capabilities are advancing rapidly, raising safety concerns and prompting government funding for cyber resilience.
Quick Take
GPT-5.5 completed a 32-step corporate network attack autonomously.
Solved a reverse-engineering challenge 70x faster than a human expert.
Researchers found a universal jailbreak bypassing safety guardrails.
UK commits £90M to cyber resilience amid rising breaches.
Market Impact Analysis
NeutralThe article focuses on AI cybersecurity, not crypto markets; no direct implication for crypto prices or adoption.
Speculation Analysis
Key Takeaways
- GPT-5.5 autonomously executed a 32-step corporate network attack in 2 out of 10 attempts.
- The model solved a reverse-engineering puzzle 70x faster than a human expert, taking just 10 minutes and costing $1.73.
- Researchers discovered a universal jailbreak that completely bypassed GPT-5.5's safety guardrails.
- The UK government committed £90 million to cyber resilience as offensive AI capabilities surge.
What Happened
The UK's AI Security Institute dropped a bombshell report: OpenAI's GPT-5.5 can autonomously pull off complex cyberattacks, placing it neck-and-neck with Anthropic's Claude Mythos. The model cracked a 12-hour reverse-engineering puzzle in just over 10 minutes and completed a simulated 32-step corporate network intrusion without human guidance. The findings signal that offensive AI is no longer theoretical—it's here, and it's improving fast.
The Numbers
On AISI's toughest test, "The Last Ones," GPT-5.5 succeeded autonomously in 2 of 10 runs, following Claude Mythos Preview which did it 3 times. The human benchmark for the same network takeover? Around 20 hours. In the reverse-engineering challenge, GPT-5.5 needed just 10 minutes 22 seconds and $1.73 in API costs. A human pro took 12 hours. Across the Expert-tier cybersecurity tasks, GPT-5.5 scored 71.4%, edging out Mythos at 68.6% and blowing past GPT-5.4 at 52.4%.
Why It Happened
The leap stems from broader AI progress. As models get better at reasoning, coding, and chaining autonomous tasks, offensive cyber capability emerges as a side effect—not a standalone breakthrough. AISI warns that if this trend holds, further leaps could come in rapid succession, blurring the line between safety testing and real-world weaponization. The same skills that make AI useful for automating workflows also make it a potent attacker.
Broader Impact
The implications extend beyond labs. A universal jailbreak that bypasses all safety guardrails raises red flags about controlling these systems. Governments are reacting: the UK pledged £90 million to bolster cyber resilience, recognizing that AI-driven attacks could soon outpace human defenses. The report adds urgency to global debates on AI regulation and the need for proactive defense measures.
What to Watch Next
- Further AISI evaluations: Watch for tests of newer models like GPT-6 or Claude's next iteration to see if offensive capabilities keep accelerating.
- Defense vs. offense race: Expect increased investment in AI-powered cyber defense tools as enterprises and governments scramble to counter autonomous threats.
- Regulatory moves: The jailbreak discovery could push governments toward stricter safety requirements or mandatory red-teaming for frontier models.
This article is for informational purposes only and does not constitute financial advice.
Always late to trends?
Join for the latest news, insights & more.
Disclaimer: Bytewit is an independent media outlet that delivers news, research, and data.
© 2026 Bytewit. All Rights Reserved. This article is for informational purposes only.