HomeMalware & ThreatsAnthropic Deems Its New Model Too Dangerous for Release

Anthropic Deems Its New Model Too Dangerous for Release

Published on

spot_img

AI-Driven Security Operations,
Artificial Intelligence & Machine Learning,
Next-Generation Technologies & Secure Development

Anthropic Limits Access to New AI Model Amid Concerns Over Misuse

Anthropic Deems Its New Model Too Dangerous for Release
Image: Shutterstock

Recently, Anthropic has declared a significant advancement in cybersecurity with the development of an artificial intelligence model that the company deems too dangerous to release to the public. This controversial decision underscores the growing unease regarding the potential misuse of advanced AI technologies.

Anthropic, a prominent player in the AI space, is already entangled in legal disputes with the U.S. federal government concerning the deployment of its AI models, particularly in realms like autonomous weapons and surveillance systems. The new model, dubbed “Claude Mythos Preview,” has reportedly identified thousands of high-severity vulnerabilities across various critical platforms, including major operating systems and web browsers.

The firm stated, “Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely. The fallout—for economies, public safety, and national security—could be severe.” This statement reflects the broader apprehensions within the technology community about AI’s potential impact on society if not appropriately governed and controlled.

In a collaborative effort to enhance cybersecurity, a coalition consisting of over 40 leading technology companies—including Microsoft, Google, and Cisco—will have access to this frontier model. The group, known as “Project Glasswing,” is allocated $100 million in usage credits to identify and address security vulnerabilities. This cooperation emphasizes a proactive strategy to fortify defenses against potential threats arising from the very technologies they continue to develop.

Cisco’s Chief Security Officer, Anthony Grieco, echoed these sentiments, highlighting the pivotal moment faced by the tech industry. “While the capabilities now available to defenders are remarkable, they soon will also become available to adversaries, defining the critical inflection point we face today,” he remarked. This statement encapsulates the pressing urgency to fortify defenses in light of evolving cybersecurity challenges.

Anthropic asserts that Mythos Preview is not merely a sophisticated fuzzer; it has successfully uncovered a 27-year-old vulnerability within OpenBSD. This vulnerability poses a severe risk, allowing an attacker to remotely crash any system running this security-focused Linux distribution, which is widely utilized in network appliances and security functions.

Furthermore, the revolutionary model has revealed vulnerabilities in the Linux kernel, permitting attackers superuser privileges by circumventing traditional security measures like kernel address space layout randomization. This discovery reflects the model’s capacity to chain multiple vulnerabilities, demonstrating its unparalleled analytical prowess. Anthropic reported instances where Mythos Preview was able to exploit two, three, or even four vulnerabilities in a single attack.

In a dedicated blog post, researchers from Anthropic elaborated on the model’s sophistication, stating that it not only identifies vulnerabilities but also comprehends the logic underpinning the code. They explained that it understands the primary function of a login feature—allowing only authorized access—even when potential bypass methods exist.

Researchers at Anthropic believe that as AI evolves, attackers and defenders will eventually reach a balance where defenders utilize these powerful tools more effectively. However, they caution that the transition could be tumultuous, with the potential for catastrophic outcomes if adversaries gain access to advanced AI models before cybersecurity professionals are adequately prepared. As a preventive measure against misuse, Anthropic has vowed to implement new safeguards to detect and mitigate malicious outputs, alongside offering upcoming recommendations addressing long-standing cybersecurity challenges. These recommendations will touch upon vital areas such as vulnerability disclosure, patch management, prioritization, and the adoption of secure-by-design practices.

Source link

Latest articles

GrafanaGhost Exploit Eases Past AI Guardrails for Silent Data Exfiltration

A recently discovered critical vulnerability, referred to as GrafanaGhost, has raised alarms in cybersecurity...

US Critical Infrastructure Vulnerable to Iranian-Linked OT Threats

CISA Reports Iranian-Linked Groups Targeting U.S. Critical Infrastructure The Cybersecurity and Infrastructure Security Agency (CISA)...

Microsoft Connects Medusa to Zero-Day Attacks

Rising Threat: Storm-1175's Rapid Medusa Ransomware Attacks In the ever-evolving landscape of cybersecurity threats, a...

What Anthropic Glasswing Indicates for the Future of Vulnerability Discovery

Anthropic Launches Project Glasswing: Transforming Cybersecurity with AI In a significant development within the cybersecurity...

More like this

GrafanaGhost Exploit Eases Past AI Guardrails for Silent Data Exfiltration

A recently discovered critical vulnerability, referred to as GrafanaGhost, has raised alarms in cybersecurity...

US Critical Infrastructure Vulnerable to Iranian-Linked OT Threats

CISA Reports Iranian-Linked Groups Targeting U.S. Critical Infrastructure The Cybersecurity and Infrastructure Security Agency (CISA)...

Microsoft Connects Medusa to Zero-Day Attacks

Rising Threat: Storm-1175's Rapid Medusa Ransomware Attacks In the ever-evolving landscape of cybersecurity threats, a...