Google Secures Gemini AI After Zero-Click Vulnerability Discovered
Overview of the GeminiJack Vulnerability
In a significant security development, Google has addressed a serious flaw in its Gemini AI tools that was capable of extracting data silently from corporate systems. Dubbed the GeminiJack vulnerability, this issue was discovered in June 2025 by researchers from Noma Security. The researchers promptly reported their findings to Google, paving the way for a critical response.
What is GeminiJack?
GeminiJack was identified as an architectural vulnerability affecting two key components of Google’s offerings: Gemini Enterprise, the suite designed for corporate AI assistance, and Vertex AI Search, which underpins AI-powered search and recommendation functionalities on Google Cloud. This vulnerability was particularly alarming because it enabled attackers to carry out indirect prompt injections.
Mechanics of the Attack
Security experts revealed that this flaw allowed malicious actors to embed harmful instructions within commonly used documents like those in Gmail or Google Docs. Once these documents were accessed, Gemini AI could unknowingly function in a way that allowed sensitive data to be extracted without the users being aware.
Key Characteristics of the Attack
Notably, the GeminiJack attack did not require any interaction from victims. Victims were not required to click on links or dismiss warnings, meaning the attack could happen seamlessly and bypass traditional corporate security measures.
How the GeminiJack Attack Worked
The researchers outlined the intricacies of the GeminiJack exploit. The attack broke down into several straightforward steps that demonstrated how attackers could leverage minimal effort for potentially devastating effects:
-
Content Poisoning: An attacker would create a seemingly innocuous Google Doc or Calendar invite, embedding hidden commands that directed Gemini Enterprise to search for sensitive terms within the corporate data accessible to it.
-
Triggering the Attack: A regular employee conducting a routine search could inadvertently activate the AI to retrieve and process the manipulated content.
-
AI Execution: Once the corrupted content was engaged, Gemini misread the hidden directives as legitimate inquiries. With its access granted by pre-existing permissions, the system then scoured the corporate workspace for the targeted sensitive data.
-
Data Exfiltration: During the AI’s response phase, it would embed a malicious image tag, which automatically transmitted the gathered information back to the attacker’s server upon being rendered in a browser. This crucial step went unnoticed, skirting normal security defenses.
The underlying cause of the vulnerability was traced to the way Gemini Enterprise’s search function operated, particularly its reliance on Retrieval-Augmented Generation (RAG). This feature allows organizations to interact with various data sources through established permissions.
Proof-of-Concept Release
Further underscoring the seriousness of the issue, Noma Security released a detailed proof-of-concept for GeminiJack on December 8, shedding light on how the exploit could be executed in real-world scenarios.
Google’s Proactive Measures
Google acknowledged receipt of the report regarding the vulnerability in August 2025 and worked collaboratively with the researchers to enact changes. Post-fix, Google updated its systems to modify how Gemini Enterprise and Vertex AI Search interacted, fully separating the two functions and enhancing their individual workflows.
Ongoing Risks and Recommendations
Even with these patches in place, security experts caution that similar forms of indirect attack could surface as organizations continue adopting AI technologies with extensive data access capabilities. They highlighted that conventional perimeter defenses and endpoint security approaches are insufficient to address scenarios where an AI assistant becomes an unwitting data exfiltration tool.
Researchers concluded that as AI systems gain broader access to corporate information and the ability to act based on user commands, the potential impact of a single security vulnerability increases substantially. They recommend that organizations should re-evaluate their trust frameworks, enhance monitoring practices, and keep abreast of developments in AI security measures.
As the landscape of AI tools evolves, continuous vigilance and proactive adaptation will be necessary to safeguard sensitive information.


