312%
Spike in AI-assisted cybercrime targeting US citizens between 2024 and 2026 -- FBI Internet Crime Complaint Center data reflecting the acceleration of AI-enabled fraud, intrusion, and exploitation across all sectors
40x
More effective than conventional cyberattacks -- CISA 2026 assessment of AI-powered offensive operations, reflecting cumulative acceleration across reconnaissance, exploitation, lateral movement, and persistence phases
$5.2M
Average cost of an AI-enhanced breach in 2026, approximately 12% higher than traditional cyberattacks -- reflecting faster lateral movement, larger blast radius, and more effective exfiltration enabled by AI-assisted operations
3-5 months
Estimated window before AI-driven exploits become the new norm in adversary operations -- Palo Alto Networks technology chief Lee Klarich, May 2026, urging organizations to act before Q3 2026

A criminal threat actor used an AI model to discover and weaponize a previously unknown authentication flaw in a popular open-source web administration tool, building a Python exploit script that would have bypassed two-factor authentication for every user of the platform in a single mass exploitation wave. Google's Threat Intelligence Group (GTIG) intercepted the AI-built zero-day exploit before the attack launched, then coordinated an emergency patch with the affected vendor.

The finding, published May 11, 2026, marks the first confirmed case of adversaries using artificial intelligence to both discover and operationalize a zero-day vulnerability in live attack infrastructure. Previous AI-assisted threat activity had involved reconnaissance and phishing automation. This incident crosses into automated exploit engineering.

The vulnerability was a logic flaw embedded in the tool's authentication flow: developers had hard-coded a trust exception that bypassed the 2FA check under specific conditions. The exploit script required no memory corruption or advanced techniques. It needed only valid user credentials plus knowledge of the hardcoded shortcut -- knowledge an LLM scanning the codebase would surface in minutes.

Google confirmed, with high confidence, that the Python script was LLM-generated. Distinctive markers included an abundance of educational docstrings explaining each function, a hallucinated CVSS score that no published advisory had assigned, and a structured, textbook Pythonic format highly characteristic of LLM training data. The threat actor's plan: deploy the script in a mass exploitation event against every reachable instance of the tool.

Every security team running open-source administration software needs to verify patch status today. Beyond that, this incident permanently reframes the threat model for AI-assisted attacks. The barrier from idea to working exploit is now measured in hours, not weeks.

How Does an AI-Built Zero-Day Exploit Actually Work?

An AI-built zero-day exploit uses a large language model to perform the work that previously required a skilled vulnerability researcher: reviewing source code, identifying logic flaws, constructing a proof-of-concept, and packaging it into a deployable script. The process compresses what once took specialists weeks into hours.

In the case GTIG documented, the targeted flaw was a semantic logic vulnerability, not a memory-corruption bug. Developers had embedded a hard-coded trust assumption into the authentication handler: under a specific code path, the application skipped the 2FA verification step when a particular session condition was present. A human auditor might spend days reading authentication logic looking for this kind of exception. An LLM prompted to review the codebase as a senior security auditor can traverse the entire code graph and flag trust assumption shortcuts in a single session.

The resulting Python script contained a structured argument parser, clean ANSI color output, and step-by-step docstrings explaining each phase of the bypass. These artifacts are consistent with LLM output but atypical for human-written exploits. Human exploit authors write terse, functional code. LLMs produce educational code with explanations, as if teaching a student the technique.

The flaw required valid credentials as a prerequisite. The attacker's operational plan involved an earlier credential acquisition phase followed by mass 2FA bypass across every reachable installation. Google assessed the threat actor was targeting a broad user population, not a single victim.

This is the operational pattern that makes AI-built zero-day exploits uniquely dangerous: AI lowers the skill floor for zero-day discovery while human operators retain full control over targeting scale.

The AI Fingerprints That Gave the Attack Away

Google identified four specific markers that attributed the exploit script to LLM generation with high confidence.

First, the script contained educational docstrings. Human exploit authors do not document their code for readability. Every function in the script included multi-line explanations written in the instructional tone of a code tutorial.

Second, the script included a hallucinated CVSS score. No published advisory had assigned a score to this vulnerability because the flaw had not yet been disclosed. The AI fabricated a plausible score and embedded it in the code comments -- a behavior characteristic of LLMs filling in expected fields rather than drawing from authoritative sources.

Third, the codebase followed textbook Pythonic structure. Clean formatting conventions, consistent import ordering, and descriptive variable names match the style patterns dominant in LLM training data. Human-written exploits in the wild are typically written for speed and disposability, not readability.

Fourth, the script included detailed help menus and clean ANSI color classes. These features improve user experience but add no exploitation capability. An LLM prompted to build a functional tool includes them automatically. A human building an attack script does not.

These fingerprints will not remain reliable detection signals for long. Adversaries who read the GTIG report will strip docstrings and introduce intentional style noise to defeat signature-based attribution. The real detection value here is confirmation that AI-generated exploit development is happening today, not as a future scenario.

Subscribe to unlock Indicators of Compromise

Free subscribers unlock full IOC lists, remediation steps, and every daily briefing.

Free daily briefing

Briefings like this, every morning before 9am.

Threat intel, active CVEs, and campaign alerts, distilled for practitioners. 50,000+ subscribers. No noise.

Nation-State Actors Using AI to Hunt Zero-Days at Scale

The criminal group behind the 2FA bypass was not the only actor observed by GTIG using AI for vulnerability exploitation. Four nation-state nexus clusters are now confirmed to be integrating AI models into their offensive research workflows at scale.

APT45 is a North Korean state-sponsored group focused on financial theft and weapons program data acquisition. GTIG analysts observed APT45 sending thousands of repetitive prompts to AI models, recursively analyzing CVEs and validating proof-of-concept exploits. The workflow functions as a force multiplier: instead of one analyst reviewing a CVE manually, APT45 runs parallel AI sessions across its full target CVE list, filtering for exploitable candidates and enriching each with automated PoC validation.

UNC2814, a China-nexus intrusion set, used a persona-driven jailbreak technique to direct an AI model to act as a senior security auditor specializing in C/C++ binary security for firmware vulnerability research. GTIG documented UNC2814 analyzing TP-Link firmware and OFTP protocol implementations using this approach, targeting the embedded device attack surface.

APT27, another China-linked group, leveraged AI tools to accelerate development of fleet management applications used to coordinate Operational Relay Box (ORB) networks -- the proxy infrastructure APT27 uses to disguise attack attribution.

Russia-nexus threat actors embedded AI-generated code in CANFAIL and LONGSTREAM malware. The AI-generated segments serve as decoy logic: non-functional code realistic enough to consume analyst time during reverse engineering. LONGSTREAM embedded 32 repetitive daylight saving time queries as camouflage.

Defenders face AI at every point in the kill chain. Read the analysis of HONESTCUE, the first malware confirmed to use live Gemini API calls for operational decisions.

AI is already accelerating vulnerability discovery, reducing the effort needed to identify, validate, and weaponize flaws. Discovery, weaponization, and exploitation are faster.

Ryan Dewhurst, watchTowr -- cited in Google GTIG report, May 2026

Subscribe to unlock Indicators of Compromise

Free subscribers unlock full IOC lists, remediation steps, and every daily briefing.

PROMPTSPY and AI-Driven Malware Operations

Beyond vulnerability discovery, GTIG documented AI integration directly inside deployed malware. PROMPTSPY is an Android backdoor that uses the Gemini 2.5 Flash Lite model as its operational decision engine. The malware serializes the Android UI hierarchy via the Accessibility API, sends it to Google's Generative Language API endpoint via HTTP POST in JSON mode, then parses the model's structured response to determine which screen elements to interact with through simulated gestures.

In practice, PROMPTSPY lets an attacker control a victim's Android device by describing objectives in natural language rather than writing hardcoded automation scripts. The malware can capture biometric authentication data and replay it, persist through Firebase Cloud Messaging, and execute commands without continuous human supervision. This is what Google researchers described as the transition from nascent AI-enabled operations to industrial-scale application of generative models within adversarial workflows.

TeamPCP, tracked as UNC6780, combined AI capabilities with supply chain compromise in March 2026. The group embedded the SANDCLOCK credential stealer inside malicious pull requests to the LiteLLM open-source project, targeting development environments that use the package to route API calls across AI providers. Extracted AWS keys and GitHub tokens were monetized through ransomware partnerships.

The supply chain angle matters: organizations that deploy AI tooling without vetting the software supply chain of those tools are expanding their attack surface. Every AI development library, API aggregator, and CLI wrapper in your environment is a potential insertion point for adversary code.

Subscribe to unlock Indicators of Compromise

Free subscribers unlock full IOC lists, remediation steps, and every daily briefing.

Who Is at Risk from AI-Built Zero-Day Exploits?

Every organization running open-source web administration tools, using AI development tooling, or relying on 2FA as a primary authentication control sits in this week's risk population. The shift is not about victim sophistication. It is about the scale at which AI enables attackers to identify and exploit vulnerabilities across all targets simultaneously.

Palo Alto Networks technology chief Lee Klarich issued a specific timeline in May 2026: "We now estimate a narrow three-to-five-month window for organizations to outpace the adversary before AI-driven exploits start to become the new norm." That window closes before Q3 2026.

Three specific risk factors elevate exposure. Organizations running open-source administration tools on internet-facing infrastructure face the highest immediate risk. The 2FA bypass Google documented was designed for mass exploitation of exactly this class of software.

Development environments that use LLM API aggregators -- including LiteLLM-based routing infrastructure -- are targets for supply chain compromise following the TeamPCP/UNC6780 attack pattern documented in March 2026.

Enterprises that rely on 2FA without monitoring for bypass conditions face an authentication control that adversaries are now specifically engineering AI to defeat. The FBI reported a 312% spike in AI-assisted cybercrime targeting US citizens between 2024 and 2026, with the average cost of an AI-enhanced breach exceeding $5.2 million.

Sector targeting mirrors existing APT priorities: financial services, critical infrastructure, defense industrial base, and organizations with access to high-value credentials or intellectual property. For a parallel view of how AI is accelerating ransomware group operations, see the SLOPOLY AI-generated malware breakdown.

How to Detect and Defend Against AI-Built Exploits

Detecting AI-built exploits before execution requires moving earlier in the attack lifecycle than traditional endpoint detection provides. By the time an exploit runs on a system, the pre-deployment detection window has closed.

Audit open-source tool authentication logic for hard-coded trust exceptions. Any code path that short-circuits 2FA under conditional logic -- particularly session-based or environment-based overrides added for developer convenience -- is a priority review target.

Monitor for abnormal outbound connections to AI provider APIs from non-developer endpoints. Legitimate Gemini API access should originate from designated development systems. Traffic from production servers, operational workstations, or managed mobile devices may indicate AI-integrated malware following the PROMPTSPY pattern.

Audit AI development libraries and API aggregators in your software supply chain. Specifically review LiteLLM, Claude-Relay-Service, CLIProxyAPI, and similar tools for unauthorized pull requests, unexpected dependency updates, or anomalous credential access events following the TeamPCP model.

Implement behavioral monitoring on 2FA systems. Log and alert on authentication attempts that succeed without completing the expected verification flow, including cases where session tokens are issued without the normal 2FA sequence.

Hunt for AI code artifacts when analyzing unknown scripts. Flag files with educational docstrings, hallucinated CVE or CVSS references, and unusually clean Pythonic formatting -- these are indicators of LLM-generated exploit code.

Block known AI aggregation tools at the network perimeter, including CLIProxyAPI, OmniRoute, and Roxy Browser, which threat actors use to pool LLM access and evade usage-limit detection.

Subscribe to unlock Remediation & Mitigation steps

Free subscribers unlock full IOC lists, remediation steps, and every daily briefing.

Why AI-Built Zero-Day Exploits Mark a New Era for Security Teams

The Google GTIG finding is not a proof-of-concept warning about a hypothetical future threat. It is a confirmed incident from this week: a working zero-day exploit was built by AI, packaged for mass deployment, and intercepted before it caused widespread damage. The next one may not be intercepted.

The strategic implication is a permanent compression of the defender's response window. When AI reduces zero-day discovery from weeks to hours and exploit packaging from days to minutes, the window between a vulnerability existing and an exploit being deployed contracts sharply. Patch cycles measured in weeks are already insufficient for this threat model.

AI-powered attacks are 40 times more effective than conventional cyberattacks, according to CISA's 2026 assessment. That figure reflects cumulative acceleration across reconnaissance, exploitation, lateral movement, and persistence -- not any single capability.

Security teams that treat AI-assisted exploitation as an emerging threat to monitor in future quarters are already behind schedule. The three-to-five-month window Palo Alto Networks describes assumes defenders act immediately. Prioritize authentication hardening, AI tooling supply chain auditing, and early-warning monitoring specifically designed to detect AI-assisted exploitation activity before the window closes.

The bottom line

The AI-built zero-day exploit Google intercepted in May 2026 proves that AI-assisted vulnerability discovery and exploit packaging have moved from theoretical risk to confirmed operational threat. Three takeaways: first, any authentication flow containing hard-coded trust exceptions is a priority audit target today; second, AI development tooling including LiteLLM and API aggregators is now a supply chain attack surface; third, the three-to-five-month window before AI-driven exploits become routine closes before Q3 2026. Audit your 2FA implementations and open-source admin tool patch status this weekend.

Frequently asked questions

What is an AI-built zero-day exploit?

An AI-built zero-day exploit is an attack tool where a large language model performs the vulnerability research and exploit coding that previously required a skilled human researcher. The AI scans source code for logic flaws, constructs a proof-of-concept, and packages it into a deployable script. Google's GTIG documented the first confirmed case in May 2026, where an LLM built a Python script that bypassed two-factor authentication in a popular open-source web administration tool.

How did hackers use AI to create the 2FA bypass zero-day?

The threat actor prompted an AI model to analyze the source code of an open-source web administration tool, acting as a security auditor. The AI identified a hard-coded trust exception in the authentication flow that allowed the 2FA check to be skipped under specific session conditions. It then generated a Python exploit script with a structured argument parser, help menus, and detailed docstrings documenting each bypass step. Google identified the script as LLM-generated through its educational formatting style, hallucinated CVSS score, and clean Pythonic structure.

Which threat actors are currently using AI for cyberattacks in 2026?

Google GTIG confirmed four nation-state nexus clusters plus criminal actors using AI in offensive operations in 2026. APT45 (North Korea) sends thousands of prompts to recursively validate exploit PoCs. UNC2814 (China) uses expert persona jailbreaks for firmware vulnerability research. APT27 (China) uses AI for ORB network tooling development. Russia-nexus actors embed AI-generated decoy code in CANFAIL and LONGSTREAM malware. An unnamed criminal group used AI to build the first documented zero-day exploit in May 2026.

Has the 2FA bypass zero-day been patched?

Yes. Google's Threat Intelligence Group coordinated responsible disclosure with the affected vendor before publishing its report on May 11, 2026. The vendor released a patch addressing the hard-coded trust exception in the authentication flow. Google also took defensive action against malware infrastructure associated with the campaign. No CVE number had been publicly assigned at the time of the GTIG report. Organizations should confirm they are running the latest version of any open-source web administration tool in their environment.

How can security teams detect AI-generated exploit code?

Current detection relies on behavioral artifacts in the code itself: educational docstrings explaining functions in tutorial language, hallucinated CVSS scores that do not correspond to published advisories, unusually clean Pythonic formatting with consistent conventions, and unnecessary user experience features such as color output and detailed help menus. These signals will degrade as adversaries learn to strip them. Network-level detection focuses on monitoring AI API traffic from non-developer endpoints and blocking known LLM aggregation tools such as CLIProxyAPI and OmniRoute.

What should organizations do right now to defend against AI-built exploits?

Audit authentication flows in all open-source admin tools for hard-coded trust exceptions that bypass 2FA. Monitor outbound connections to AI provider APIs from production systems, which may indicate PROMPTSPY-style malware. Review the software supply chain of all AI development libraries including LiteLLM. Implement session-level logging that alerts when 2FA completion is absent from an authenticated session record. Block AI aggregation proxy tools at network egress points. Apply all vendor patches from the past 30 days immediately, prioritizing authentication components.

What is APT45 and how is it using AI for cyberattacks?

APT45 is a North Korean state-sponsored threat group focused on financial theft and acquiring data related to weapons programs. Google GTIG documented APT45 sending thousands of repetitive prompts to AI models to recursively analyze CVE disclosures and validate proof-of-concept exploits at scale. This workflow allows the group to maintain a larger arsenal of validated exploit capabilities than would be practical through manual research alone, functioning as an AI-powered force multiplier for its vulnerability operations.

What is PROMPTSPY malware and how does it use AI?

PROMPTSPY is an Android backdoor that uses the Gemini 2.5 Flash Lite AI model as its operational decision engine. The malware serializes the Android device UI hierarchy via the Accessibility API and sends it to Google's Generative Language API endpoint. The AI model returns structured instructions for simulated gestures such as clicks and swipes, allowing an attacker to control the device by describing objectives in natural language. PROMPTSPY can capture biometric authentication data for replay attacks and persists via Firebase Cloud Messaging.

Sources & references

  1. Google Cloud Blog — Adversaries Leverage AI for Vulnerability Exploitation, Augmented Operations, and Initial Access
  2. The Hacker News — Hackers Used AI to Develop First Known Zero-Day 2FA Bypass for Mass Exploitation
  3. SecurityWeek — Google Detects First AI-Generated Zero-Day Exploit
  4. CNBC — Palo Alto: AI-driven cyberattacks will start to be the new norm in months
  5. BleepingComputer — Google: Hackers Used AI to Develop Zero-Day Exploit for Web Admin Tool

Free resources

25
Free download

Critical CVE Reference Card 2025–2026

25 actively exploited vulnerabilities with CVSS scores, exploit status, and patch availability. Print it, pin it, share it with your SOC team.

No spam. Unsubscribe anytime.

Free download

Ransomware Incident Response Playbook

Step-by-step 24-hour IR checklist covering detection, containment, eradication, and recovery. Built for SOC teams, IR leads, and CISOs.

No spam. Unsubscribe anytime.

Free newsletter

Get threat intel before your inbox does.

50,000+ security professionals read Decryption Digest for early warnings on zero-days, ransomware, and nation-state campaigns. Free, weekly, no spam.

Unsubscribe anytime. We never sell your data.

Eric Bang
Author

Founder & Cybersecurity Evangelist, Decryption Digest

Cybersecurity professional with expertise in threat intelligence, vulnerability research, and enterprise security. Covers zero-days, ransomware, and nation-state operations for 50,000+ security professionals weekly.

Free Brief

The Mythos Brief is free.

AI that finds 27-year-old zero-days. What it means for your security program.

Joins Decryption Digest. Unsubscribe anytime.

Daily Briefing

Get briefings like this every morning

Actionable threat intelligence for working practitioners. Free. No spam. Trusted by 50,000+ SOC analysts, CISOs, and security engineers.

Unsubscribe anytime.

Mythos Brief

Anthropic's AI finds zero-days your scanners miss.