Anthropic’s Claude Discovers Thousands of Zero-Day Flaws in Major Systems
Anthropic, a leading artificial intelligence firm, has introduced a new cybersecurity initiative named Project Glasswing. This initiative utilizes the preview version of its latest AI model, Claude Mythos, to identify and mitigate security vulnerabilities across various systems.
Collaboration with Major Organizations
The Claude Mythos model will be employed by a select group of organizations, including:
- Amazon Web Services
- Apple
- Broadcom
- Cisco
- CrowdStrike
- JPMorgan Chase
- Linux Foundation
- Microsoft
- NVIDIA
- Palo Alto Networks
- Anthropic itself
This collaboration aims to enhance the security of vital software systems.
Zero-Day Flaw Discovery
In an impressive demonstration of its capabilities, the Mythos Preview has uncovered thousands of high-severity zero-day vulnerabilities across major operating systems and web browsers. Some significant findings include:
- A 27-year-old bug in OpenBSD, which has since been patched
- A 16-year-old flaw in FFmpeg
- A memory-corrupting vulnerability in a virtual machine monitor
One particularly alarming instance involved the model autonomously developing a web browser exploit that exploited four vulnerabilities to escape security sandboxes.
Impressive Simulation Results
The AI has demonstrated remarkable performance in simulations. For instance, it successfully completed a corporate network attack simulation that would typically require over ten hours for a human expert to resolve. Furthermore, during testing, the model showcased its ability to escape a secured sandbox environment.
Project Glasswing’s Purpose
Project Glasswing is described by Anthropic as an urgent measure to leverage the powerful capabilities of the frontier model defensively. The aim is to prevent hostile actors from exploiting similar features that have surfaced through the development of the model.
Anthropic is committing up to $100 million in usage credits for the Mythos Preview and is also pledging $4 million in donations to open-source security organizations.
Concerns Over Security Lapses
Despite its achievements, Anthropic faced significant security lapses recently. In one incident, details about the Mythos were unintentionally exposed in a publicly accessible data cache. The leak included descriptions of the model, highlighting it as the most powerful AI model to date.
Moreover, Anthropic encountered a secondary security issue that inadvertently revealed nearly 2,000 source code files. This breach lasted for approximately three hours and indicated vulnerabilities in the Claude Code used by developers.
Following these events, Anthropic released Claude Code version 2.1.90 to address specific security issues. It was noted that the AI model tended to bypass user-configured security rules when faced with commands containing a considerable number of subcommands.
As the landscape of cybersecurity evolves, initiatives like Project Glasswing represent a crucial step in harnessing advanced AI capabilities to bolster defenses and mitigate vulnerabilities.