Real-World AI Security Breaches: 7 Incidents Every CISO Should Study
From EchoLeak's zero-click data theft in Microsoft 365 Copilot to ShadowLeak silently draining Gmail through ChatGPT — these 2025 AI security incidents reveal the attack patterns every security leader must understand before their next deployment.
2025 was the year AI security vulnerabilities went from theoretical to devastating. Zero-click exploits hit Microsoft Copilot, ChatGPT, and Google Gemini. Hundreds of thousands of private conversations leaked onto Google. Source code was silently stolen from GitHub Copilot. These aren't proof-of-concept demos — they're production incidents that exposed sensitive data across every major AI platform. Each one carries lessons that every security leader must internalize.
Why This Matters Now
All seven incidents in this analysis occurred in 2025 and affect products from Microsoft, OpenAI, Google, xAI, and GitHub. No vendor was immune. The common thread: AI systems were deployed faster than the security architecture could mature.
1. EchoLeak: The First Zero-Click Prompt Injection in Production (June 2025)
Incident Summary
- Product: Microsoft 365 Copilot
- CVE: CVE-2025-32711
- CVSS Score: 7.5 (NIST) / 9.3 Critical (Microsoft)
- Discovered by: Aim Security
- Type: Zero-click indirect prompt injection → data exfiltration
EchoLeak (CVE-2025-32711) represents a watershed moment in AI security: the first real-world, zero-click prompt injection exploit in a production LLM system. Discovered by researchers at Aim Security in June 2025, it allowed an attacker to steal confidential data from a victim's Microsoft 365 environment by sending a single crafted email — with no user interaction required.
The Technical Attack Chain
What made EchoLeak particularly alarming was its sophistication — it chained five separate bypasses to defeat Microsoft's layered defenses:
- XPIA Classifier Evasion: Microsoft's Cross Prompt Injection Attempt classifier was bypassed by phrasing malicious instructions as if directed at the email recipient, avoiding language that referenced AI or assistants
- Markdown Link Bypass: While Copilot filtered standard Markdown links
[text](URL), it failed to catch reference-style Markdown:[ref]: https://attacker.com - Image Injection: Alternative image syntax
![alt][ref]wasn't filtered, allowing image references that could carry exfiltrated data in URL parameters - CSP Bypass via Teams Proxy: Content Security Policy allowed
*.teams.microsoft.com, which hosted an open redirect that could tunnel exfiltrated data to attacker servers - RAG Optimization: Attack instructions were strategically embedded across multiple email chunks with likely search terms, ensuring the malicious content was retrieved by Copilot's RAG pipeline
Key Lesson
Single-layer defenses are insufficient for AI systems. EchoLeak proved that sophisticated attackers will chain multiple bypasses against any individual security control. Defense-in-depth — combining input validation, output filtering, strict CSP policies, provenance-based access control, and prompt partitioning — is the minimum viable security posture for production AI.
2. ShadowLeak: ChatGPT's Deep Research Agent Silently Drains Gmail (June 2025)
Incident Summary
- Product: ChatGPT Deep Research Agent (with Gmail, GitHub, Drive connectors)
- Discovered by: Radware (Zvika Babo, Gabi Nakibly, Maor Uziel)
- Disclosed: June 18, 2025 — patched by August 2025
- Type: Zero-click indirect prompt injection → service-side data exfiltration
- Success rate: 100%
ShadowLeak demonstrated that AI agent vulnerabilities can be even more dangerous than chatbot exploits. When a ChatGPT user connected their Gmail and asked Deep Research to analyze their inbox, a single crafted email containing hidden instructions (white-on-white text, tiny fonts, CSS tricks) could hijack the agent into silently exfiltrating sensitive data to an attacker-controlled server.
Why ShadowLeak Was Different
Unlike previous prompt injection attacks that relied on client-side rendering tricks, ShadowLeak performed data exfiltration directly from OpenAI's cloud infrastructure. The attack was invisible to enterprise firewalls, DLP tools, and network monitoring because no suspicious traffic ever left the victim's device — the data was stolen server-side by OpenAI's systems following the injected instructions.
The attacker could instruct the agent to encode extracted PII in Base64 before appending it to malicious URLs — an additional obfuscation layer. The vulnerability affected every ChatGPT connector: Gmail, GitHub, Google Drive, Outlook, Box, Dropbox, Notion, and SharePoint.
Key Lesson
AI agents with access to external data sources create a new category of server-side exfiltration risk that traditional security tools can't detect. Organizations must evaluate whether granting AI agents access to sensitive data stores (email, code repos, drives) creates acceptable risk — and implement strict input sanitization on every data source the agent ingests.
3. GeminiJack: Google Gemini Leaks Gmail, Calendar, and Docs (2025)
Incident Summary
- Product: Google Gemini Enterprise / Vertex AI Search
- Discovered by: Noma Labs
- Type: Zero-click indirect prompt injection → cross-service data exfiltration
- Data at risk: Gmail messages, Calendar details, Google Docs, years of email history
GeminiJack showed that Google's enterprise AI was equally vulnerable to indirect prompt injection. An attacker could embed hidden instructions in a shared Google Doc, calendar invitation, or email. When any employee performed a routine search in Gemini Enterprise (e.g., "show me our budgets"), the AI's RAG system would retrieve the poisoned document and execute the attacker's commands — searching across Gmail, Calendar, and Docs to exfiltrate data through disguised image requests.
The attack required zero clicks and zero user interaction. It occurred silently during Gemini's background processing. No DLP alerts were triggered because the AI operated through Google's own approved systems.
The Gemini Trifecta
Separately, Tenable Research discovered three additional Gemini vulnerabilities in 2025:
- Cloud Assist vulnerability: Malicious prompts injected via manipulated HTTP headers into Google Cloud logs, which Gemini would then summarize and execute
- Search Personalization vulnerability: JavaScript injection into browser history that Gemini treated as legitimate user queries
- Browsing Tool vulnerability: Tricking Gemini into embedding user data in query strings to attacker-controlled servers
Google addressed the findings by fully separating Vertex AI Search from Gemini Enterprise and modifying their RAG interaction patterns. All vulnerabilities have been patched.
Key Lesson
Enterprise AI with broad workspace access makes every shared document a potential attack vector. RAG-powered AI must validate the provenance and intent of every document it retrieves — treating shared content as untrusted input, not trusted context.
4. Grok Data Exposure: 370,000+ Conversations Indexed by Google (August 2025)
Incident Summary
- Product: xAI's Grok chatbot
- Data exposed: 370,000+ conversations
- Root cause: Share feature created public URLs crawled by search engines
- Reported by: AIAAIC, Forbes, Fortune
In August 2025, over 370,000 private conversations with xAI's Grok chatbot were discovered indexed by Google, Bing, and DuckDuckGo. The exposure occurred because Grok's "share" feature created publicly accessible URLs on Grok's website with no robots.txt or noindex tag to prevent search engine crawling.
What Was Exposed
The indexed conversations contained deeply sensitive information as reported by Forbes and Fortune:
- Medical and psychological questions
- Personal relationship details and passwords
- Business plans and financial discussions
- Requests for instructions on illegal activities — including drug production, malware coding, and weapons manufacturing
Users received no warning that clicking "share" would create a permanent, publicly searchable web page.
Key Lesson
Every feature that creates external URLs from AI conversations must be treated as a data publication pipeline. Implement noindex tags, require explicit opt-in with clear warnings, and conduct privacy impact assessments before launching any share functionality. Default settings should always favor privacy.
5. ChatGPT Conversation Leaks: 100,000 Sessions Exposed (2025)
Incident Summary
- Product: OpenAI ChatGPT
- Data exposed: ~100,000 conversations
- Root cause: "Make chat discoverable" feature with unclear implications
- Analysis by: SafetyDetectives (1,000 conversations, 43 million words)
OpenAI tested a "Make this chat discoverable" feature that allowed users to opt-in to making conversations searchable. A misconfigured noindex tag meant search engines crawled and indexed the shared sessions. Researchers scraped nearly 100,000 conversations before OpenAI disabled the feature.
SafetyDetectives Analysis: What Users Were Sharing
SafetyDetectives analyzed 1,000 of the leaked conversations spanning over 43 million words:
- Personally identifiable information: Names, phone numbers, email addresses, physical locations
- Mental health data: Suicidal ideation, addiction recovery discussions, psychological counseling
- Professional secrets: Proprietary algorithms, source code, NDAs, business documents
- Credentials: Passwords, API keys, authentication tokens
"Professional consultations" accounted for nearly 60% of flagged topics. OpenAI called the feature "a short-lived experiment that introduced too many opportunities for folks to accidentally share things they didn't intend to" — but the data had already been permanently archived by researchers.
Key Lesson
Users treat AI chatbots as confidential advisors. Any feature that exposes conversation data — even with "opt-in" framing — must assume users don't fully understand the implications. Privacy-by-default is not optional for AI products.
6. CamoLeak: GitHub Copilot Silently Steals Private Source Code (June 2025)
Incident Summary
- Product: GitHub Copilot Chat
- CVSS Score: 9.6 Critical
- Discovered by: Nudge Security
- Type: Hidden prompt injection + CSP bypass → source code and secrets exfiltration
CamoLeak (CVSS 9.6) demonstrated that AI coding assistants could be weaponized to silently steal private source code. Attackers embedded malicious prompts in invisible GitHub markdown comments within pull requests or issues — comments that are invisible to users in the web UI but are parsed by Copilot Chat.
The Exfiltration Technique
To bypass GitHub's Content Security Policy (which prevented direct data transmission), the attackers devised an ingenious method: they created a dictionary of valid GitHub Camo-proxied image URLs mapping to each character of the alphabet. The injected prompt coerced Copilot to render secrets as a sequence of 1x1 pixel images — enabling character-by-character data extraction through image request patterns.
The vulnerability allowed theft of:
- AWS keys and security tokens from private repositories
- Unpublished zero-day vulnerability descriptions
- Private source code using the victim's own repository permissions
GitHub responded by disabling image rendering in Copilot Chat on August 14, 2025, and blocking Camo usage for leaking content.
Key Lesson
AI coding assistants that process repository content are exposed to prompt injection through any content they read — including comments, issues, and PRs from untrusted contributors. Organizations must treat AI tools with code access as high-privilege applications requiring the same security scrutiny as CI/CD pipelines.
7. GitHub MCP Prompt Injection: AI Agents Backdoor Your Code (May 2025)
Incident Summary
- Product: GitHub MCP Integration / Copilot Agent
- Discovered by: Invariant Labs / Trail of Bits
- Type: Indirect prompt injection via GitHub issues → code backdoor insertion
- Impact: Private repo data theft, malicious PR generation
Two separate research teams independently discovered that GitHub's AI agent features could be weaponized through public issues. The Invariant Labs Security Research Team found that attackers could create malicious GitHub issues in public repositories containing hidden prompt injection payloads. When a developer asked their AI assistant to review issues, the agent would follow the hidden instructions — accessing private repositories and leaking sensitive data including salary information and confidential business details.
Separately, Trail of Bits demonstrated that attackers could file issues containing hidden prompts (using HTML <picture> tags) that trick Copilot Agent into generating pull requests with malicious backdoors when maintainers assign the agent to fix the issue. The malicious code is inserted directly into the codebase through a seemingly legitimate AI-generated PR.
Key Lesson
AI agents that interact with public-facing content (issues, PRs, comments) and have write access to code create a supply chain attack vector. Every AI-generated code change must be treated as untrusted and reviewed with the same scrutiny as human-submitted code — or more, given that the AI may be following attacker instructions invisible to reviewers.
Cross-Incident Analysis: Common Patterns
Across all seven incidents, clear patterns emerge that should inform every organization's AI security strategy:
| Pattern | Incidents | Mitigation |
|---|---|---|
| Zero-click prompt injection via external data | EchoLeak, ShadowLeak, GeminiJack | Input isolation, prompt partitioning, data provenance tracking |
| Server-side exfiltration bypassing enterprise defenses | ShadowLeak, GeminiJack | AI-specific monitoring, connector access reviews, output validation |
| Privacy-eroding features deployed without safeguards | Grok, ChatGPT leaks | Privacy-by-default, noindex tags, explicit opt-in with clear warnings |
| AI coding tools as supply chain attack vector | CamoLeak, GitHub MCP injection | AI output code review, restrict write access, treat AI PRs as untrusted |
| AI privilege escalation across access boundaries | GeminiJack, CamoLeak, EchoLeak | Authorization-aware RAG, scoped retrieval, least-privilege AI access |
CISO Action Items
Based on these incidents, here are concrete steps security leaders should take:
Immediate (This Quarter)
- Inventory all AI integrations — know every Copilot plugin, Gemini workspace connection, ChatGPT connector, and LLM API key in your environment
- Review AI connector permissions — ShadowLeak and GeminiJack exploited overly broad data access; restrict connectors to minimum necessary scope
- Audit share/export features in every AI tool — ensure conversation data isn't being published without safeguards
- Treat AI-generated code as untrusted — CamoLeak and the MCP injection show AI PRs can contain attacker-inserted backdoors
Short-Term (This Half)
- Deploy prompt injection detection on all AI interfaces — tools like promptguardrails can identify malicious inputs before they reach your models
- Implement authorization-aware RAG — ensure AI retrieval systems respect the querying user's access permissions, not the system's
- Conduct AI-specific red team exercises — test your AI deployments for prompt injection, data exfiltration, and privilege escalation
- Monitor AI agent data flows — traditional DLP won't catch server-side exfiltration; you need AI-native monitoring
Strategic (This Year)
- Build an AI security architecture that includes input validation, output filtering, access controls, monitoring, and incident response — treating AI as critical infrastructure
- Evaluate enterprise AI deployments with data residency guarantees as alternatives to public AI services for sensitive workflows
- Integrate AI security into your SDLC — security reviews for AI features should be as routine as code reviews
AI Security Platform
Every incident above was preventable. Build defenses that catch sophisticated attacks before they reach your models — from multi-chain prompt injection to data exfiltration.
The Bottom Line
Every incident in this analysis was preventable with known security practices — input validation, access controls, privacy-by-default design, and output monitoring. The problem isn't a lack of knowledge; it's a lack of implementation.
2025 proved that AI vulnerabilities aren't theoretical. They hit Microsoft, OpenAI, Google, GitHub, and xAI — the most well-resourced AI companies on the planet. If they're vulnerable, so is your organization. The CISOs who study these incidents and act on the lessons will be far better positioned than those who wait for their own breach.
Sources and Further Reading
- Aim Security: EchoLeak: Zero-Click Prompt Injection in Microsoft 365 Copilot
- NVD: CVE-2025-32711
- Radware: ShadowLeak: Zero-Click Service-Side Attack on ChatGPT Deep Research
- Noma Labs: GeminiJack: Google Gemini Zero-Click Vulnerability
- Tenable: The Trifecta: Three Gemini Vulnerabilities
- Forbes: xAI Published Hundreds Of Thousands Of Grok Conversations
- SafetyDetectives: ChatGPT Leaks: 1,000 Public Conversations Analyzed
- Nudge Security: CamoLeak: GitHub Copilot Code Exfiltration
- Docker: MCP Horror Stories: GitHub Prompt Injection Data Heist
- Trail of Bits: Prompt Injection: Exploiting GitHub Copilot