Defend your organization from business scams
Progress 0 / 12
Chapters (12)
Deepfakes and AI-enhanced attacks
AI technology has revolutionized cybercrime, enabling attackers to create convincing deepfake audio, video, and text at scale. The 194% surge in AI fraud and 3,000% increase in deepfakes represent a paradigm shift in social engineering sophistication.
The Scale of the Problem
2024 Statistics:
- 194% surge in AI-enabled fraud
- 3,000% increase in deepfake incidents
- $1 trillion projected cost globally by 2027
- Voice cloning: 3 seconds of audio = 85% match
- Real-time deepfake video now possible
- 40% of BEC attacks use AI-generated content
AI Attack Capabilities
Voice cloning:
- 3-second audio sample sufficient
- 85%+ accuracy in matching tone
- Real-time conversation possible
- Can clone any voice from public recordings
Video deepfakes:
- Real-time video manipulation
- Face swapping on video calls
- Lip-syncing to match fake audio
- High quality from consumer hardware
Text generation:
- Perfect grammar phishing emails
- Context-aware responses
- Personality mimicry
- Multi-language capability
Image generation:
- Fake IDs and documents
- Profile photos for fake personas
- Manipulated screenshots
- Realistic but fraudulent evidence
Real-World Cases
Arup $25M deepfake (2024):
- Video call with fake CFO and executives
- Real-time deepfakes of multiple people
- Finance employee authorized 15 transactions
- Sophisticated AI orchestration
Voice cloning CEO fraud:
- AI-cloned CEO voice calling CFO
- Requested urgent wire transfer
- Perfect voice match fooled recipient
- Stopped only by verification procedures
Detection Challenges
Why deepfakes are hard to detect:
- Quality improving exponentially
- Real-time generation now possible
- Detection tools lag behind creation tools
- Human senses insufficient
- Context and situation matter more than tech
Subtle indicators:
- Slight audio delays or glitches
- Unnatural eye movement or blinking
- Inconsistent lighting or shadows
- Background artifacts
- Emotional expression timing off
- But these are disappearing rapidly
Verification Procedures
For voice calls:
- Ask personal questions only real person knows
- Request callback on known number
- Use challenge-response code words
- Verify through separate channel
- Listen for unnatural pauses or glitches
For video calls:
- Ask person to perform specific actions
- Request they hold up item with today’s date
- Ask unexpected questions
- Switch to in-person for high-stakes decisions
- Use multi-person verification
For all high-risk requests:
- Out-of-band verification mandatory
- Multiple verification methods
- Don’t rely solely on seeing/hearing
- Context matters (why this request, why now)
Protection Strategies
Technical defenses:
- Deepfake detection tools (limited effectiveness)
- Multi-factor authentication
- Digital signatures for communications
- Recorded verification procedures
- AI-powered anomaly detection
Procedural defenses:
- Verification protocols that can’t be bypassed
- Challenge questions changed regularly
- Code words for sensitive operations
- Multi-person approval for large transactions
- Waiting periods prevent real-time manipulation
Cultural defenses:
- Awareness that deepfakes exist and are good
- Permission to verify even CEO
- “Trust but verify” as default
- Reporting suspected deepfakes encouraged
The Arms Race
Attacker advantages:
- AI tools democratized (easy to use)
- Quality improving monthly
- Real-time generation achieved
- Detection harder than creation
Defender strategies:
- Process over technology
- Multiple verification layers
- Human judgment enhanced by tech
- Assume compromise possible
- Build verification into culture
Key Takeaways
- ✅ 194% surge in AI-enabled fraud attacks
- ✅ Voice cloning from 3 seconds of audio
- ✅ Real-time deepfakes now possible
- ✅ Technology detection insufficient - process matters
- ✅ Out-of-band verification mandatory for high-risk requests
- ✅ Challenge questions and code words essential
- ✅ Assume seeing/hearing isn’t enough - always verify
- ✅ Build culture where verification is expected, not questioned