$2.77B
In AI-powered BEC losses across 21,442 incidents in 2026, per FBI IC3 reporting
$25M
Lost by Arup in a single deepfake video call where every participant, including the CFO, was AI-generated
3 sec
Of audio required to clone a voice convincingly using modern tools like ElevenLabs and open-source alternatives
30%
Of enterprises will find standalone identity verification unreliable by 2026, per Gartner prediction

Deepfake fraud has moved from a theoretical threat to the fastest-growing category of financial crime targeting enterprises. The attack is simple in concept and devastating in execution: AI synthesizes a convincing replica of a trusted executive's voice or video, an employee receives what appears to be a direct instruction from their CEO or CFO, and money moves before anyone questions it.

The Arup incident established a new upper bound for individual deepfake fraud losses. A finance employee joined a video call that included a convincing AI-generated version of the company's CFO and several other colleagues, all fabricated. They authorized multiple wire transfers totaling $25 million. The call looked real, the voices sounded real, and the video appeared real. The attack succeeded entirely because the employee had no out-of-band verification mechanism.

This guide covers the technical detection capabilities, procedural controls, and organizational defenses that stop deepfake fraud. Most of the effective controls are procedural rather than technical, which means they can be implemented immediately without waiting for vendor procurement.

How Deepfake Attacks Work in 2026

Understanding the attack mechanics is prerequisite to building effective defenses. Deepfake fraud in 2026 uses three delivery methods, each with different detection requirements.

Voice cloning via phone or audio call is the most common and most scalable attack. Modern voice synthesis tools including ElevenLabs, Resemble AI, and open-source alternatives can produce highly convincing voice replicas from as little as three seconds of training audio. That audio is freely available for most executives: earnings calls, conference presentations, YouTube interviews, and LinkedIn video posts all provide sufficient training material. The attacker calls a finance team member, employee, or administrator, speaks as the executive, and requests an urgent wire transfer, credential reset, or policy exception.

Synthetic video calls are higher effort but higher conversion. The attacker creates a deepfake video avatar of the executive and joins a video call using that avatar. The Arup incident used this technique. Video deepfakes have visual artifacts that trained observers can detect, but untrained employees in a routine video call context do not scrutinize facial movement or lip sync carefully enough to notice.

Hybrid attacks combine AI-generated voice or video with compromised email accounts for follow-up. An attacker who has compromised a vendor's email account sends a payment request, then follows up with a deepfake phone call from the CEO confirming the request urgency. The two-channel corroboration makes the attack far more convincing than either channel alone.

Voice-only cloning attacks

Attacker synthesizes executive voice from public audio and calls finance or IT staff with urgent requests. Lowest effort, highest scale, most common.

Deepfake video calls

AI-generated video avatar joins a video conference. Higher effort but defeats visual authentication. The Arup $25M attack used this method.

Hybrid email plus voice attacks

Compromised or spoofed email sets up the scenario, deepfake voice call provides executive confirmation. Two-channel corroboration significantly increases victim compliance.

AI-enhanced vishing

Real-time voice conversion transforms an attacker's live voice into the target executive's voice during a phone call, enabling two-way conversations with the synthesized voice.

Procedural Controls: The Most Effective Defense

Technical deepfake detection is improving but is not yet reliable enough to serve as the primary defense. Procedural controls that require out-of-band verification for high-risk actions are the most robust defense available right now and can be implemented without any technology investment.

The single most effective control is a callback verification requirement for all wire transfers, credential resets, and sensitive data access requests received via phone or video call, regardless of how convincing the requester appears. The callback must use a pre-registered phone number from your corporate directory, not a number provided by the caller. Even a $25 million deepfake attack fails if the finance employee calls the real CFO on their known mobile number before executing the transfer.

Code word protocols add a second layer. Agree on a rotating verification code word shared among executive leadership and key finance and IT staff. Any request involving financial authorization or privileged access requires the requester to provide the current code word. A deepfake caller who does not know the code word cannot pass verification regardless of how convincing the voice or video appears. Rotate code words monthly and distribute them through a secure channel separate from email.

Approval workflow requirements eliminate single-point authorization. Wire transfers above a defined threshold should require approval from two distinct individuals through a system-enforced workflow, not a chain of verbal confirmations. An attacker who convincingly impersonates one executive cannot simultaneously impersonate a second executive in a different channel without a dramatically more complex operation.

Emergency urgency as a red flag: the most consistent characteristic of deepfake fraud attempts is artificial urgency. Requests that bypass normal approval workflows because of time pressure should automatically trigger additional verification, not expedited approval. Train finance and executive assistants that urgency from an executive is a reason to verify more carefully, not less.

Free daily briefing

Briefings like this, every morning before 9am.

Threat intel, active CVEs, and campaign alerts, distilled for practitioners. 50,000+ subscribers. No noise.

Technical Detection Capabilities

Technical deepfake detection is an active research area with commercial products emerging in 2025 and 2026. Understanding their capabilities and limitations helps set realistic expectations for what technology can and cannot catch.

Audio deepfake detection analyzes voice recordings for statistical artifacts introduced by synthesis models: unnatural pitch variation, synthesis-model-specific spectral signatures, and inconsistencies in background noise patterns. Pindrop, Nuance, and specialized AI detection vendors offer real-time voice authentication that can flag synthesized audio during phone calls. Detection accuracy varies significantly by synthesis tool and audio quality; detection rates against state-of-the-art synthesis remain below 90% in independent testing.

Video deepfake detection looks for visual artifacts in facial rendering: unnatural blinking patterns, lip-sync inconsistencies, boundary artifacts around facial features, and lighting inconsistencies between the face and background. Intel's FakeCatcher and Microsoft's Video Authenticator use pixel-level analysis to detect synthesis artifacts. Practical limitations include: detection accuracy degrades significantly with compressed video (as in most video calling platforms); real-time detection latency may exceed what is practical during a live call; and detection models trained on 2024 synthesis techniques may miss 2026 synthesis improvements.

A practical technical deployment is post-call analysis for high-stakes video calls: record calls involving financial authorization requests and run them through detection tools after the fact. This does not prevent the initial attack but enables rapid detection and reversal before wire transfers settle, which typically takes one to two business days for international transfers.

Security Awareness Training for Deepfake Fraud

Deepfake fraud defeats the employee's primary trust signal: the familiar voice and face of a known executive. Standard security awareness training that teaches employees to be skeptical of unknown senders or suspicious links does not address this threat directly. Deepfake-specific training requires updating the mental model employees use to authenticate executive requests.

The core behavioral change is separating the apparent identity of the requester from the authorization to approve a request. Employees should understand that voice and video can now be convincingly synthesized, which means appearing to be the CEO on a call is no longer sufficient authorization for a high-risk action. The authorization comes from the verified approval workflow, not from the person's apparent identity.

Simulated deepfake exercises run controlled deepfake attack simulations against finance, HR, and IT staff, similar to phishing simulations. Vendors including Aware and DeepMedia offer synthetic voice simulation exercises. The goal is not to embarrass employees who are fooled, but to build muscle memory for the out-of-band verification response before a real attack tests it under urgency pressure.

Executive digital footprint awareness matters because executives whose voice and video are widely available online are the highest-risk impersonation targets. Brief executives on minimizing unnecessary public audio and video while continuing legitimate media presence, and on the existence of the deepfake threat so they understand why employees may ask for code word verification even in situations where the executive knows their own identity is genuine.

Financial Controls That Limit Deepfake Fraud Impact

Even when detection fails, financial controls can limit the damage. The most effective financial controls create time delays and multi-party checkpoints that allow detection before funds are irreversible.

Wire transfer hold periods for first-time payees or large transaction amounts give fraud teams time to verify before funds clear. Many banks offer configurable hold periods as a business account feature. A 24-hour hold on any first-time wire transfer above $50,000 provides a detection window for deepfake fraud without materially affecting normal business operations.

Dual authorization requirements in banking platforms enforce two-person approval at the financial institution level, not just in internal approval workflows. This means that even if an attacker successfully manipulates one authorized signer through deepfake fraud, the transfer cannot complete without a second signer's independent approval in the banking system.

Payment limits and change controls require multi-party approval to modify wire transfer limits, add new payees, or change payment instructions. A deepfake attack that requests an emergency increase in wire transfer limits before executing the fraud requires compromising two separate approval chains rather than one.

Cyber insurance policy review is warranted specifically for deepfake fraud coverage. Many commercial crime policies written before 2024 do not explicitly cover AI-synthetic social engineering losses. Review your policy language and work with your broker to ensure deepfake-induced wire fraud is covered before you need to file a claim.

The bottom line

Deepfake fraud cannot be reliably stopped by asking employees to listen carefully to whether a voice sounds natural or to watch for visual glitches in a video call. The controls that work are procedural: mandatory out-of-band callback verification using pre-registered numbers for all high-risk requests, code word protocols that cannot be replicated by a caller who was not present when the code was distributed, and multi-party financial approval workflows that require two independent authorizations before funds move. Implement these before a deepfake call tests your organization's defenses under real urgency pressure.

Frequently asked questions

What is deepfake CEO fraud and how is it different from regular BEC?

Traditional business email compromise (BEC) impersonates executives through spoofed or compromised email addresses. Deepfake CEO fraud adds AI-synthesized voice or video to impersonate the executive directly in a phone call or video call. This defeats the safeguard that employees use when they are skeptical of email requests: calling the person to confirm. When the person who answers the phone also sounds exactly like the executive, the employee's verification check is defeated.

How little audio does an attacker need to clone someone's voice?

Modern commercial voice synthesis tools can produce convincing voice replicas from as little as three seconds of audio. Most executives have abundant public audio available through earnings calls, conference panel recordings, YouTube interviews, and LinkedIn video posts. The attacker does not need to compromise any internal systems to obtain training audio; public sources are sufficient for most senior executives at publicly visible organizations.

Can deepfake detection technology reliably catch these attacks?

Not yet reliably enough to serve as the primary defense. Current detection tools achieve accuracy rates below 90% against state-of-the-art synthesis in independent testing, and accuracy degrades further with compressed video as used by video calling platforms. Detection technology is improving but is best used as a supplementary layer alongside procedural controls, not as a replacement for out-of-band verification requirements.

What is a callback verification protocol and how do we implement it?

A callback verification protocol requires that any request received via phone or video call involving a high-risk action (wire transfer, credential reset, sensitive data access) be verified by calling the requester back on a pre-registered number from your corporate directory, not a number provided during the call. The implementation is: define the high-risk action categories, document pre-registered numbers for all executives in a secure internal directory, train relevant staff on the procedure, and require no exceptions regardless of urgency claimed by the original caller.

Should executives avoid appearing in public video and audio to reduce deepfake risk?

Minimizing unnecessary exposure is sensible, but avoiding all public audio and video is not realistic for most executives and not necessary for effective defense. The procedural controls (callback verification, code words, multi-party approval) stop deepfake fraud independent of whether the attacker has high-quality training audio. Brief executives on the threat and make them aware that employees may request code word verification, but do not restrict legitimate media presence on the basis of deepfake risk alone.

What should we do if we suspect we have already been victimized by deepfake fraud?

Act within the wire transfer settlement window. International wire transfers typically take one to two business days to settle. Contact your bank immediately to initiate a wire recall. File a complaint with the FBI IC3 (Internet Crime Complaint Center) and your local FBI field office, as they have wire fraud recovery mechanisms including the Financial Fraud Kill Chain. Contact your cyber insurance carrier. Preserve all evidence including call recordings, email chains, and system logs for forensic investigation and insurance claim purposes.

How do we train employees to verify executive requests without offending the executive?

Frame callback verification as a compliance requirement, not personal distrust. Employees should be trained to say: 'I need to follow our verification procedure before I can process this. Let me call you back on your registered number.' Any real executive who understands the deepfake threat environment will recognize this as the correct response. Executives should be briefed on the verification protocol and asked to actively support it, including acknowledging to their teams that they should always verify unusual requests even from the CEO.

Sources & references

  1. CybelAngel: Voice Cloning Is the New BEC
  2. Brightside AI: Deepfake CEO Fraud $50M Voice Cloning Threat
  3. Gartner: 30% of Enterprises Will Find Identity Verification Unreliable by 2026
  4. Cogent: Deepfake Onslaught 2026 Enterprise Defense
  5. FBI IC3: Business Email Compromise Report 2025

Free resources

25
Free download

Critical CVE Reference Card 2025–2026

25 actively exploited vulnerabilities with CVSS scores, exploit status, and patch availability. Print it, pin it, share it with your SOC team.

No spam. Unsubscribe anytime.

Free download

Ransomware Incident Response Playbook

Step-by-step 24-hour IR checklist covering detection, containment, eradication, and recovery. Built for SOC teams, IR leads, and CISOs.

No spam. Unsubscribe anytime.

Free newsletter

Get threat intel before your inbox does.

50,000+ security professionals read Decryption Digest for early warnings on zero-days, ransomware, and nation-state campaigns. Free, weekly, no spam.

Unsubscribe anytime. We never sell your data.

Eric Bang
Author

Founder & Cybersecurity Evangelist, Decryption Digest

Cybersecurity professional with expertise in threat intelligence, vulnerability research, and enterprise security. Covers zero-days, ransomware, and nation-state operations for 50,000+ security professionals weekly.

Free Brief

The Mythos Brief is free.

AI that finds 27-year-old zero-days. What it means for your security program.

Joins Decryption Digest. Unsubscribe anytime.

Daily Briefing

Get briefings like this every morning

Actionable threat intelligence for working practitioners. Free. No spam. Trusted by 50,000+ SOC analysts, CISOs, and security engineers.

Unsubscribe anytime.

Mythos Brief

Anthropic's AI finds zero-days your scanners miss.