← Back to Scams & phishing

Defend your organization from business scams

Deepfakes and AI-enhanced attacks

AI technology has revolutionized cybercrime, enabling attackers to create convincing deepfake audio, video, and text at scale. The 194% surge in AI fraud and 3,000% increase in deepfakes represent a paradigm shift in social engineering sophistication.

The Scale of the Problem

2024 Statistics:

  • 194% surge in AI-enabled fraud
  • 3,000% increase in deepfake incidents
  • $1 trillion projected cost globally by 2027
  • Voice cloning: 3 seconds of audio = 85% match
  • Real-time deepfake video now possible
  • 40% of BEC attacks use AI-generated content

AI Attack Capabilities

Voice cloning:

  • 3-second audio sample sufficient
  • 85%+ accuracy in matching tone
  • Real-time conversation possible
  • Can clone any voice from public recordings

Video deepfakes:

  • Real-time video manipulation
  • Face swapping on video calls
  • Lip-syncing to match fake audio
  • High quality from consumer hardware

Text generation:

  • Perfect grammar phishing emails
  • Context-aware responses
  • Personality mimicry
  • Multi-language capability

Image generation:

  • Fake IDs and documents
  • Profile photos for fake personas
  • Manipulated screenshots
  • Realistic but fraudulent evidence

Real-World Cases

Arup $25M deepfake (2024):

  • Video call with fake CFO and executives
  • Real-time deepfakes of multiple people
  • Finance employee authorized 15 transactions
  • Sophisticated AI orchestration

Voice cloning CEO fraud:

  • AI-cloned CEO voice calling CFO
  • Requested urgent wire transfer
  • Perfect voice match fooled recipient
  • Stopped only by verification procedures

Detection Challenges

Why deepfakes are hard to detect:

  • Quality improving exponentially
  • Real-time generation now possible
  • Detection tools lag behind creation tools
  • Human senses insufficient
  • Context and situation matter more than tech

Subtle indicators:

  • Slight audio delays or glitches
  • Unnatural eye movement or blinking
  • Inconsistent lighting or shadows
  • Background artifacts
  • Emotional expression timing off
  • But these are disappearing rapidly

Verification Procedures

For voice calls:

  • Ask personal questions only real person knows
  • Request callback on known number
  • Use challenge-response code words
  • Verify through separate channel
  • Listen for unnatural pauses or glitches

For video calls:

  • Ask person to perform specific actions
  • Request they hold up item with today’s date
  • Ask unexpected questions
  • Switch to in-person for high-stakes decisions
  • Use multi-person verification

For all high-risk requests:

  • Out-of-band verification mandatory
  • Multiple verification methods
  • Don’t rely solely on seeing/hearing
  • Context matters (why this request, why now)

Protection Strategies

Technical defenses:

  • Deepfake detection tools (limited effectiveness)
  • Multi-factor authentication
  • Digital signatures for communications
  • Recorded verification procedures
  • AI-powered anomaly detection

Procedural defenses:

  • Verification protocols that can’t be bypassed
  • Challenge questions changed regularly
  • Code words for sensitive operations
  • Multi-person approval for large transactions
  • Waiting periods prevent real-time manipulation

Cultural defenses:

  • Awareness that deepfakes exist and are good
  • Permission to verify even CEO
  • “Trust but verify” as default
  • Reporting suspected deepfakes encouraged

The Arms Race

Attacker advantages:

  • AI tools democratized (easy to use)
  • Quality improving monthly
  • Real-time generation achieved
  • Detection harder than creation

Defender strategies:

  • Process over technology
  • Multiple verification layers
  • Human judgment enhanced by tech
  • Assume compromise possible
  • Build verification into culture

Key Takeaways

  • 194% surge in AI-enabled fraud attacks
  • Voice cloning from 3 seconds of audio
  • Real-time deepfakes now possible
  • Technology detection insufficient - process matters
  • Out-of-band verification mandatory for high-risk requests
  • Challenge questions and code words essential
  • Assume seeing/hearing isn’t enough - always verify
  • Build culture where verification is expected, not questioned
Author:
How To Use Internet
Last updated:
11/30/2025