Google recently introduced its new Integrated Development Environment (IDE) called Antigravity, but security experts have quickly shown that its features can be misused.
A report has come out detailing how attackers can exploit a vulnerability known as indirect prompt injection to steal sensitive information by manipulating the very AI agents that are meant to enhance productivity.
The security research firm PromptArmor discovered that this flaw takes advantage of Antigravity’s default configurations.
By embedding hidden instructions in tiny text on a webpage, hackers can trick the AI into ignoring file protection measures and extract confidential data to a publicly accessible logging site.

Despite the seriousness of this issue, Google considers these actions to be part of the tool’s intended functionality, which leaves corporate codebases vulnerable. This situation underscores a significant disconnect between how these “reasoning” models are marketed and their actual security capabilities.
The attack process begins with a compromised web source, like a third-party guide that contains harmful instructions.
Attackers conceal the prompt injection in a font size so small that it is invisible to human eyes but can be read by the Gemini model. Once the AI processes this malicious input, it is instructed to gather sensitive information from the user’s local files.
Initially, Gemini recognizes that the .env file is protected because it is included in the .gitignore list, and it refuses to access it due to safety protocols.
However, the AI’s autonomy allows it to bypass this restriction. Following the attacker’s instructions, it uses the terminal command cat to view the file contents, which demonstrates how easily the AI can override its own safety measures.
After gaining access to the sensitive data, the agent encodes this information into a URL format. The final step in the attack involves sending the stolen data to webhook.site, a service that logs incoming requests.
This exfiltration is made possible because webhook.site is mistakenly included in Antigravity’s default list of permitted URLs, which undermines network security by allowing data to leave without triggering alerts.
PromptArmor decided to disclose these findings sooner than the usual 90-day period, citing Google’s previous classification of similar security issues as “intended behavior.”
The researchers stated that Google was already aware of the risks of data theft highlighted by their research, so they felt responsible disclosure was unnecessary.
There is a fundamental disagreement between security experts and Google regarding what constitutes an acceptable level of risk for these autonomous tools.
The documentation from Google backs up PromptArmor’s findings, as their Bug Hunters platform clearly states that the Antigravity agent has access to files and can execute commands, classifying these as invalid report types.
This “won’t fix” approach conflicts with the messaging surrounding the launch of the Antigravity IDE, where executives promoted Gemini 3 Pro as an advanced reasoning engine capable of solving complex problems.
Additional research has confirmed these risks, with another security expert identifying remote command execution vulnerabilities that extend beyond just browser attacks.
PromptArmor noted that the vulnerabilities are not limited to specific setups, finding three more ways that data could be stolen without relying on the browser features being active.
Industry leaders are currently facing the challenge of balancing the seamless experience promised by “agent-first” development with the need for strict security measures. Antigravity is designed with a default policy that lets the agent make decisions without human input, which significantly reduces oversight.
Adding to the concern is a policy that allows the agent to run system commands like cat or curl automatically, without needing user approval.
These design choices prioritize speed over security, creating an environment that is highly vulnerable to exploitation.
Experts refer to this collection of risks as a “Lethal Trifecta.” The vulnerabilities arise from the agent having access to untrusted web input, private data from the codebase, and the ability to communicate externally over the internet. When all three factors are present, the chances of data theft become nearly certain without proper isolation measures.
While other tools like Cursor and Windsurf share similar risks, Antigravity’s lenient default settings make it particularly prone to immediate data breaches. Typically, other platforms require explicit user consent for network requests to new domains, which adds an extra layer of protection.
To mitigate these risks, experts suggest that agents operating in “YOLO-mode,” which disables all safety checks, should be run in isolated environments like firewalled Virtual Machines instead of directly on the main operating system.
Without such precautions, these findings could deter businesses from adopting autonomous coding tools until stricter network controls and sandboxing measures are put in place by developers.
Other Stories You May Like