AI Firms Tackle Prompt Injection

AI Firms Tackle Prompt Injection

2 November 2025

Tech giants Google DeepMind, Anthropic, and Microsoft are intensifying efforts to defend against 'indirect prompt injection attacks'. This emerging security flaw allows hackers to embed malicious instructions within data sources that AI systems access, such as documents or web pages. When the AI processes this compromised content, it can be manipulated into performing unintended actions, including data leaks, spreading misinformation, or executing malicious code.

Unlike direct prompt injection, where attackers directly input malicious prompts, indirect attacks exploit the AI's interaction with external data. The AI unknowingly treats the embedded commands as legitimate instructions, bypassing traditional security measures. This poses a significant risk to AI-powered applications, as it can lead to unauthorised access and privilege escalation.

To counter this threat, tech companies are developing multi-layered defences, including data governance strategies, content sanitisation, and real-time threat detection. These measures aim to distinguish between legitimate content and malicious instructions, ensuring the safe and reliable operation of AI systems.

Source:ft.com

AI generated content may differ from the original.

Published on 2 November 2025
artificialintelligenceintelligenceaisecuritypromptinjectiondeepmindmicrosoft
  • AI: Data Privacy Paradox

    AI: Data Privacy Paradox

    Read more about AI: Data Privacy Paradox
  • AI Fuels Expense Fraud Surge

    AI Fuels Expense Fraud Surge

    Read more about AI Fuels Expense Fraud Surge
  • AI Capex: Risky Business?

    AI Capex: Risky Business?

    Read more about AI Capex: Risky Business?
  • OpenAI Faces Public Interest Test

    OpenAI Faces Public Interest Test

    Read more about OpenAI Faces Public Interest Test