HomeCII/OTThe Skeleton Key Reveals Malicious Content

The Skeleton Key Reveals Malicious Content

Published on

spot_img

Microsoft has raised an alarm about a new kind of attack called “Skeleton Key” that enables users to bypass the safety measures incorporated into generative AI models like ChatGPT. This prompt injection attack manipulates the context around typically prohibited chatbot requests, enabling users to access offensive, harmful, or illegal content.

Initially, most commercial chatbots would reject requests for instructions on creating dangerous malware capable of disrupting power plants. However, by framing the request as being for educational purposes within a safe research context and providing a disclaimer, the AI may proceed to provide the requested information without censorship.

Microsoft’s Chief Technology Officer for Azure, Mark Russinovich, explained that once the guardrails are disregarded, AI models struggle to differentiate between malicious or unsanctioned requests and those with legitimate intentions. This loophole has been termed as the Skeleton Key technique due to its ability to completely bypass security measures and disclose the full extent of the model’s knowledge.

Multiple generative AI models, including those managed by Microsoft Azure, Meta, Google Gemini, Open AI, Mistral, Anthropic, and Cohere, were found to be susceptible to this technique. Microsoft promptly addressed the issue by implementing prompt shields in Azure to detect and block this tactic, along with software updates to enhance security.

Although Microsoft resolved the vulnerability in its Azure platform, other vendors are advised to implement necessary fixes. Microsoft also provided recommendations for administrators to safeguard their AI models against prompt injection attacks, such as input filtering to identify harmful intents, an additional guardrail to prevent safety instruction tampering, and output filtering to block responses that breach safety protocols.

This caution from Microsoft underscores the evolving nature of cybersecurity threats, especially in the realm of AI technologies. As advancements in AI continue to shape various industries, it is crucial for organizations to stay vigilant and implement robust security measures to mitigate risks associated with prompt injection attacks like Skeleton Key.

Source link

Latest articles

Trizetto Alerts 3.4 Million About 2024 Hack Detected in 2025

Compromise Affects Healthcare Clients of TriZetto’s Revenue Cycle Management Services In a significant incident revealing...

Check Point Launches Secure AI Advisory Service for Enterprise AI Governance

Check Point Software has introduced an innovative Secure AI Advisory Service, aimed at assisting...

Tycoon 2FA Explodes as Europol and Vendors Dismantle Phishing Platform

Phishing-as-a-Service Platform Gains Popularity Among Cyber Threat Actors for Bypassing Multifactor Authentication In the ever-evolving...

Cisco releases urgent patches for severe firewall vulnerabilities

Cisco Security Flaws: An Overview of Recent Vulnerabilities Recent reports have surfaced regarding serious vulnerabilities...

More like this

Trizetto Alerts 3.4 Million About 2024 Hack Detected in 2025

Compromise Affects Healthcare Clients of TriZetto’s Revenue Cycle Management Services In a significant incident revealing...

Check Point Launches Secure AI Advisory Service for Enterprise AI Governance

Check Point Software has introduced an innovative Secure AI Advisory Service, aimed at assisting...

Tycoon 2FA Explodes as Europol and Vendors Dismantle Phishing Platform

Phishing-as-a-Service Platform Gains Popularity Among Cyber Threat Actors for Bypassing Multifactor Authentication In the ever-evolving...