What is Claude Mythos and what risks does it pose?

TL;DR
Anthropic's new AI model, Claude Mythos, reportedly outperforms humans in certain hacking tasks, raising concerns among regulators and financial institutions. The tool's capabilities are under scrutiny, with access granted to select tech companies through Project Glasswing.
Key points
- Claude Mythos is a new AI model by Anthropic.
- It reportedly outperforms humans in hacking tasks.
- Regulators are concerned about its risks to digital services.
- Project Glasswing allows select tech companies access to Mythos.
- Mythos competes with OpenAI's ChatGPT and Google's Gemini.
Mentioned in this story
In recent weeks, the AI world has been a-buzz following claims made by leading firm, Anthropic, regarding its new model, Claude Mythos.
The company says it found the tool can outperform humans at some hacking and cyber-security tasks, which has prompted discussions by regulators, legislators and financial institutions about the dangers it could pose to digital services.
Several tech giants have been given access to Mythos via an initiative called Project Glasswing, designed to strengthen resilience to Mythos itself.
But others point out that it is in Anthropic's interests to suggest its tool has never-seen-before capabilities, meaning - as ever with AI - the job of distinguishing between justified claims and hype can be tricky.
Mythos is one of Anthropic's latest models developed as part of its broader AI system called Claude. It encompasses the company's AI assistant and family of models, rivalling OpenAI's ChatGPT and Google's Gemini.
It was revealed by Anthropic in early April as "Mythos Preview".
Researchers who test how AI models handle particular requests or tasks, known as "red-teams", said in a report Mythos was "strikingly capable at computer security tasks".
They found the tool could locate dormant bugs lurking in decades-old code and easily exploit them.
So rather than make it widely available to Claude users, Anthropic gave 12 tech companies access via Project Glasswing, which it described as "an effort to secure the world's most critical software".
They include cloud computing giant Amazon Web Services, device manufacturers Apple, Microsoft and Google, and chip-makers Nvidia and Broadcom.
Crowdstrike, whose faulty software update caused a major global outage in July 2024, is also among the project's partners, with Anthropic saying it has also given access to Mythos to more than 40 organisations responsible for critical software.
Anthropic says during tests it found the model was highly skilled at cyber-security and hacking tasks, outperforming humans.
"Mythos Preview has already found thousands of high-severity vulnerabilities, including some in every major operating system and web browser," Anthropic claimed on 7 April.
"Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely."
It said it could locate - without much oversight - critical bugs in need of immediate action in old systems, including one vulnerability which had been present in a system for 27 years, and suggest ways to exploit them.
Some finance ministers, central bankers and financiers have since expressed serious concerns about it, fearing the model could undermine the security of financial systems.
Canadian finance minister François-Philippe Champagne told the BBC Mythos had been discussed at a International Monetary Fund (IMF) meeting in Washington DC this week.
"Certainly it is serious enough to warrant the attention of all the finance ministers," he said, describing the tech as an "unknown unknown".
Bank of England boss Andrew Bailey told the BBC "we are having to look very carefully now what this latest AI development could mean for the risk of cyber crime."
Meanwhile, the EU has said it is also in discussions with Anthropic about its concerns around Mythos.
Ciaran Martin, former head of the UK's National Cyber Security Centre, told the BBC earlier this week the claim Mythos could unearth critical vulnerabilities much more quickly than other AI models had "really shaken people".
"The second thing is that even with existing weaknesses that we know about, but organisations might not have patched against, might not be well defended against, it's just a really good hacker," he said.
Many independent cyber-security analysts and experts have not yet been able to test it themselves and some remain sceptical about Mythos' performance.
The UK's AI Safety Institute recently concluded that while a very powerful model, its biggest threat would be against poorly defended, vulnerable systems.
"We cannot say for sure whether Mythos Preview would be able to attack well-defended systems," its researchers said.
So where there is good cybersecurity, this model would, in theory, hopefully be stopped.
Fears relating to AI are nothing new.
New models and tools are coming out all the time, and are often accompanied by promises to revolutionise our lives, for better or worse.
Capitalising on this mix of fear and excitement over AI and its future impact has also become a hallmark of the sector and its marketing strategies in recent years.
In the case of Mythos, we still do not know enough about to know whether these hopes or fears are justified, or more a reflection of the hype surrounding the industry.
In either cases, according to the NSCS, the most important thing we can do now is not panic and instead focus on the need to get the basic cyber-security right.
After all, most hackers do not need super AI tools to breach systems when much simpler attacks often suffice.
"For some this is an apocalyptic event, for others it seems to be a lot of hype," Martin told the BBC.
But he said whether it was this tool or subsequent ones made by Anthropic or its rivals, alongside the risk there was an opportunity to build a safer online world.
"In the medium-term, there's an opportunity to use these tools to fix a lot of the underlying vulnerabilities in the internet," he said.
Sign up for our Tech Decoded newsletter to follow the world's top tech stories and trends. Outside the UK? Sign up here.
Q&A
What is Claude Mythos and how does it work?
Claude Mythos is an AI model developed by Anthropic, designed to perform tasks related to hacking and cybersecurity, potentially outperforming human capabilities.
What risks does Claude Mythos pose to digital services?
The risks include potential misuse in hacking and cybersecurity threats, prompting discussions among regulators and financial institutions about its implications for digital security.
Who has access to Claude Mythos through Project Glasswing?
Several tech giants have been granted access to Claude Mythos via Project Glasswing, an initiative aimed at enhancing resilience against the model's capabilities.
How does Claude Mythos compare to other AI models like ChatGPT and Gemini?
Claude Mythos is part of Anthropic's broader AI system and is positioned as a competitor to models like OpenAI's ChatGPT and Google's Gemini, focusing on advanced tasks in cybersecurity.





