Hackers tricked ChatGPT into raiding Gmail inboxes, researchers warn

Hackers briefly weaponized ChatGPT to steal data straight from Gmail inboxes.

Security researchers at Radware revealed how they tricked OpenAI’s Deep Research tool into exfiltrating personal emails using a stealthy “prompt injection” exploit. The loophole is now closed, but it underscores how easily agentic AI can be hijacked — with victims often unaware.

Key Takeaways

  • ChatGPT’s Deep Research tricked into leaking Gmail data, Radware says.
  • Attack used prompt injection hidden in plain sight.
  • Exploit ran on OpenAI’s cloud, bypassing standard defenses.
  • OpenAI patched the flaw in June after disclosure.
  • Outlook, GitHub, Google Drive may face similar risks.

Radware researchers discovered a prompt injection attack that tricked ChatGPT’s Deep Research tool into exfiltrating sensitive Gmail data. The exploit, dubbed “Shadow Leak,” bypassed traditional defenses by running on OpenAI’s cloud. OpenAI patched the flaw in June, but experts warn other connected services remain at risk.

Shadow Leak: How Hackers Hijacked ChatGPT

Cybersecurity firm Radware has revealed how it manipulated OpenAI’s Deep Research agent inside ChatGPT to pull sensitive information directly from Gmail inboxes. The exploit, called Shadow Leak, highlights the growing risks tied to so-called agentic AI — tools designed to act autonomously on users’ behalf.

Deep Research, launched earlier this year, can access connected apps like Gmail, calendars, and cloud drives when authorized. Radware researchers showed that a malicious email containing hidden instructions could quietly turn the assistant into a data thief.

How the Attack Worked

The trick relied on a prompt injection — hidden text that looks harmless to humans but instructs the AI to behave differently.

In this case, Radware hid the exploit inside an email. When the user later employed Deep Research to analyze their inbox, the tool stumbled across the hidden prompt. It was instructed to search for HR emails and personal details and send them outside the system.

Unlike traditional phishing, the user would see nothing suspicious. Even worse, because the operation executed within OpenAI’s own infrastructure, standard security tools had no visibility.

A Rollercoaster of Failures and Breakthroughs

Radware researchers admitted the exploit took significant trial and error. “This process was a rollercoaster of failed attempts, frustrating roadblocks, and, finally, a breakthrough,” they wrote.

But once refined, the technique worked — demonstrating how easily attackers could flip an AI agent into a double agent.

Risks Beyond Gmail

While the proof-of-concept targeted Gmail, Radware warned that other connectors could be vulnerable. Services like Outlook, Google Drive, Dropbox, and GitHub — often used for sensitive business data — could be exploited the same way.

That means contracts, meeting notes, or even source code might be exposed if attackers refine these methods.

Industry Response

OpenAI plugged the vulnerability in June after Radware disclosed its findings. The company has not publicly detailed its patch, but said it continues to invest in safeguarding agentic AI.

Security analysts note, however, that prompt injection attacks are inherently difficult to prevent. Because the instructions are embedded in natural language, there’s no simple filter to catch every malicious attempt.

Why It Matters

The Shadow Leak case underlines a key problem: users may never know when they’ve been compromised. Unlike a phishing link or malware infection, the exploit leaves no obvious trace.

It also raises broader questions about the safety of outsourcing tasks to AI agents. These tools can boost productivity, but as attackers learn to weaponize them, the potential fallout grows.

The Bigger Picture

Prompt injection attacks aren’t new. Researchers and hackers alike have demonstrated that they can be exploited to rig peer reviews, manipulate chatbots into scams, or even control smart home devices.

But what makes Shadow Leak different is that it operated within OpenAI’s cloud, sidestepping firewalls and endpoint defenses. That blind spot could prove attractive to threat actors.

What Happens Next

Experts expect AI security to become a front-line concern for enterprises adopting agentic AI. Some recommend limiting the access these tools have and monitoring logs for unusual activity.

Regulators, particularly in the EU, are also likely to scrutinize whether providers of AI assistants are transparent about risks.

Conclusion

ChatGPT’s Shadow Leak exploit may be patched, but it won’t be the last. As AI agents become more powerful, the very features that make them useful also make them dangerous. The challenge now is finding safeguards before real attackers strike.

Source

Also Read..

Leave a Comment