tech

April 15, 2026

Anthropic, Google, and Microsoft paid AI agent bug bounties, then kept quiet about the flaws

Security researcher Aonan Guan hijacked AI agents from Anthropic, Google, and Microsoft via prompt injection attacks on their GitHub Actions integrations, stealing API keys and tokens in each case. All three companies paid bug bounties quietly, $100 from Anthropic, $500 from GitHub, an undisclosed amount from Google, but none published public advisories or assigned CVEs, leaving users on older versions unaware of the risk.

Anthropic, Google, and Microsoft paid AI agent bug bounties, then kept quiet about the flaws

TL;DR

  • AI agents from Anthropic, Google, and Microsoft are vulnerable to prompt injection attacks that can steal API keys and tokens.
  • Researcher Aonan Guan successfully exploited vulnerabilities in Anthropic's Claude Code Security Review, Google's Gemini CLI Action, and GitHub's Copilot Agent.
  • The companies paid bug bounties but did not issue public advisories or CVEs for the vulnerabilities.
  • The attacks exploit the inability of AI models to reliably distinguish between trusted data and malicious instructions.
  • The lack of disclosure means users running older versions of these AI integrations may remain unaware of the security risks.
  • The issue is a structural problem in AI agent design, not an isolated bug, with similar vulnerabilities found in other AI tools and agent marketplaces.
  • There is no established framework for disclosing AI agent vulnerabilities, unlike traditional software bugs.

Continue reading the original article

Made withNostr