Skip to content

Mythos AI: Anthropic Probes Unauthorized Access as Pentagon Disputes ‘Kill Switch’ Control

Pentagon with a large red button. Mythos AI: Anthropic Probes Unauthorized Access as Pentagon Disputes 'Kill Switch' Control

A dual crisis has erupted at Anthropic due to its advanced Mythos AI model, and there is a potential security breach as well as an ongoing lawsuit with the Pentagon surrounding control of the AI for military purposes. The company is currently investigating whether or not an unauthorized group accessed Mythos via a third party vendor’s environment; however, they have stated that nothing has been uncovered, and there is no evidence that malicious actors have obtained the model.

Mythos AI: Anthropic Probes Unauthorized Access as Pentagon Disputes 'Kill Switch' Control: The powerful cybersecurity model, deemed too dangerous for public release, has triggered global alarm and a legal battle over military use.
Source: United States Court of Appeals for the District of Columbia Circuit

The Mythos AI Dilemma

The model referred to as Mythos AI, developed and built by Anthropic, can be used as a type of dangerous tool for public use (a frontier cybersecurity tool) because of the risk of identifying, developing, and finding vulnerabilities in critical infrastructure software at a rapid pace. Mythos has only been provided to 11 entities within the U.S. and only one in the UK. Due to the restricted use of Mythos, it has prompted emergency responses from several Central Banks and Intelligence Organizations globally. The Governor of the Bank of England stated Mythos can “open the entire cyber risk industry,” and a Russian organization stated it is “worse than a nuclear weapon.”

The U.S. Pentagon and Anthropic are also in a legal dispute in Federal Court, where the Pentagon has classified Anthropic as a ‘Supply Chain Risk’, and Anthropic claims that it has no “kill switch”; therefore, they have no visibility, no technical ability, or control of the AI technology developed and created once deployed by the military. To this point, the Pentagon argues that Anthropic is wrongfully intervening in the way its technology can be used in conjunction with sensitive operations, while the AI firm argues that the Pentagon, for instance, can test the technology before deployment. Anthropic has also stated that its current usage policy prohibits the use of autonomous weapon systems and the use of mass surveillance.

Mythos AI: Anthropic Probes Unauthorized Access as Pentagon Disputes 'Kill Switch' Control: The powerful cybersecurity model, deemed too dangerous for public release, has triggered global alarm and a legal battle over military use.
Benchmark experiment for Mythos Preview, which developed working exploits 181 times, and achieved register control on 29 more. (Source: Anthropic Red Team)

What is Happening Next

As of right now, the investigation of claims of unauthorized access is still ongoing, while the legal battle over the government’s contracts with the Pentagon continues with a split decision on access to provide any new work with Anthropic for the military, though it continues to be associated with other government agencies. 

Moreover, Anthropic is now working on Project Glasswing, which was recently introduced. It claims to be “an urgent initiative to help secure the world’s most critical software.” By using Mythos Preview, the AI firm has already found thousands of high-severity vulnerabilities, aiming to turn this tool into “invaluable for defensive work.”

At the same time, the ongoing global discussion around who controls the strongest AI and how to keep them from being used inappropriately has shifted from a theoretical concept to a pressing and imminent reality. 

It is well known that in the coming years, AI tech will overtake most of the current work humans perform in different industries. We are seeing that now with massive layoffs and a control-machine switch. The future could be overwhelming dystopian. Going further, do we need government-based regulations for everything? Do we need more control over our everyday lives? Just as an example, even the so-called decentralized finance (DeFi) is losing the little decentralization base it was created upon, and this is a serious concern in the crypto space. How much control are we able to give up when we cannot control ourselves most of the time?

Final Take

Anthropic created an extremely powerful AI that it couldn't release, and then it was potentially accessed by an unknown party. Now, the same company tells a judge it doesn't have a way to take that AI away from the Pentagon after it's given to them. Whether it be called irony or inevitability, the conclusion seems to be that we are at the beginning of a new age in which we are now breaking things we cannot fix. And that should keep everyone awake at night.

Disclaimer: All content provided on Times Crypto is for informational purposes only and does not constitute financial or trading advice. Trading and investing involve risk and may result in financial loss. We strongly recommend consulting a licensed financial advisor before making any investment decisions.

A Web3 Journalist at TimesCrypto with a knack for turning complex ideas into engaging stories. With a solid Tech background, Alan has led teams to create and refine impactful projects across industries, working in firms such as IBM, Cisco Systems, and Telecom. He’s passionate about Blockchain, Finance, Science, bringing a unique blend of technical expertise and creative flair to every piece he writes. When he’s not crafting content, you’ll find him diving deep into research or just having some fun!

Zoomable Image