ЦиберСецурити СЕЕ

Confused Pilot Targets AI Systems with Data Poisoning in New Attack

Researchers at the University of Texas at Austin’s SPARK Lab have identified a new cyber-attack method known as ConfusedPilot, which targets Retrieval-Augmented Generation (RAG) based AI systems like Microsoft 365 Copilot. Led by Professor Mohit Tiwari, CEO of Symmetry Systems, the team discovered how attackers could manipulate AI-generated responses by injecting malicious content into documents referenced by the AI.

This manipulation could potentially lead to misinformation and flawed decision-making within organizations, posing a significant threat as 65% of Fortune 500 companies are either currently using or planning to implement RAG-based systems. The ConfusedPilot attack method is alarming as it only requires basic access to a target’s environment and can persist even after the malicious content has been removed. Additionally, the researchers found that the attack could bypass existing AI security measures, causing concern across various industries.

The way ConfusedPilot works involves several steps. First, the attacker poisons the data environment by adding specially crafted content to documents indexed by the AI system. When a query is made, the AI references the tainted document, potentially misinterpreting the malicious content as instructions and generating misinformation or falsely attributing its response to credible sources. Even after the tainted document is removed, the corrupted information may linger in the system, making it a persistent threat.

Large enterprises using RAG-based AI systems are particularly vulnerable to such attacks, especially those that rely on multiple user data sources. This increases the risk of manipulation as the AI can be influenced by seemingly harmless documents added by insiders or external partners. Stephen Kowski, field CTO at SlashNext, emphasizes the risks associated with making decisions based on inaccurate or incomplete data, highlighting the potential consequences such as missed opportunities, revenue loss, and reputational damage.

To mitigate the risks posed by ConfusedPilot, the researchers recommend implementing certain strategies. These include implementing data access controls to limit who can upload or modify documents referenced by AI systems, conducting regular data audits to ensure data integrity, segmenting sensitive information to prevent the spread of compromised data, using AI security tools to monitor outputs for anomalies, and ensuring human oversight of AI-generated content before making critical decisions.

Amit Zimerman, co-founder and chief product officer at Oasis Security, emphasizes the importance of evaluating the effectiveness of AI-enabled security tools in specific contexts. Rather than being swayed by marketing claims, organizations should test these tools against real-world data to ensure they provide actionable insights and uncover previously unseen threats.

In conclusion, the discovery of the ConfusedPilot attack method highlights the importance of bolstering cybersecurity measures, especially when deploying AI systems that can be vulnerable to manipulation. By adopting the recommended mitigation strategies and staying vigilant against emerging threats, organizations can better protect themselves from cyber-attacks targeting AI systems.

Извор линк

Exit mobile version