Back to Fraud Forward

AI Fraud Threats: Agentic Attacks and the Future of Fraud

September 17, 2025
Hailey Windham
HOST
Fraud Forward, Sardine
Karisse Hendrick
Award-Winning Cyberfraud Expert
Listen on
YouTube
Apple Podcasts
Spotify
Pocket casts
Overcast
Share
Share this episode

What is up fraud fighters, and welcome to Fraud Forward!

Today’s episode was recorded live in Austin, and it’s a special one because I got to sit down again with someone many of you already know and respect in this space.

Karisse Hendrick.

Karisse is an award winning cyberfraud expert and the host of Fraudology, and every time we talk we end up digging into where fraud is heading next.

And right now, that conversation is all about AI fraud threats.

Because AI fraud threats are no longer hypothetical conversations happening at conferences.

They’re happening right now.

Fraudsters are already using artificial intelligence to automate scams, accelerate attacks, and test defenses faster than traditional fraud detection systems were designed to respond.

In this episode, we talk about how AI driven fraud is changing the structure of fraud attacks and why fraud teams have to rethink how they monitor risk in an AI powered environment.

What you’ll hear in this episode

  • How AI fraud threats are changing the fraud landscape in real time
  • Why agentic AI attacks are creating autonomous fraud attacks at machine speed
  • Examples of platform exploitation fraud and AI misuse by fraudsters
  • The growing challenge of AI enabled social engineering
  • Why fraud prevention in the AI era requires faster collaboration and stronger visibility

You should listen to this episode if

  • You lead fraud prevention or cyberfraud programs
  • Your team is evaluating emerging fraud threats tied to artificial intelligence
  • You are responsible for fraud leadership and AI risk governance
  • You want to understand adaptive fraud attacks and autonomous fraud behavior
  • You are preparing your organization for the future of fraud prevention

If you liked this episode, be sure to subscribe and review the podcast on iTunes, Spotify, YouTube, or wherever you listen. It helps more fraud fighters find these conversations.

Episode notes & key takeaways

Before we double click on the notes, I just want to say that my marketing team told me I need to structure these notes a certain way in order for people to find my podcast. The below is a bit of that 😀

AI Fraud Threats Are Already Reshaping Active Attack Environments

AI fraud threats are not theoretical anymore.

Fraud teams are encountering AI driven fraud in active environments where attacks move faster, adapt quicker, and operate across multiple platforms simultaneously.

Fraudsters are no longer relying only on static phishing scripts or predictable social engineering.

AI powered scams allow attackers to iterate quickly, test defenses, and adjust tactics in real time.

That compresses the detection window dramatically.

Fraud programs that rely on slow escalation processes or retrospective analysis will struggle to keep up with emerging fraud threats powered by artificial intelligence.

Agentic AI Attacks Are Changing Fraud Execution

Agentic AI attacks represent one of the most important developments in cyberfraud and AI risk.

Unlike simple automation, autonomous fraud attacks can operate with conditional logic and adaptive behavior once deployed.

These systems can:

  • Test fraud detection controls across platforms
  • Modify tactics when friction appears
  • Scale attacks across multiple systems simultaneously
  • Continue operating without constant human oversight

This creates a new category of AI fraud detection challenges.

Traditional fraud detection models built around predictable behavior struggle against adaptive fraud attacks.

Fraud prevention in the AI era requires systems that respond as quickly as attacks evolve.

Platform Exploitation Fraud Expands the Attack Surface

AI platforms are becoming part of the fraud attack surface.

When fraudsters exploit AI powered systems, platform exploitation fraud spreads through environments that users already trust.

For example, attackers may hijack AI tools to distribute malicious links, generate scam content, or amplify social engineering campaigns.

This creates layered risk for institutions.

Fraud risk in AI systems depends not only on external attackers but also on how platforms design, deploy, and monitor their AI capabilities.

Organizations must evaluate how AI misuse by fraudsters could impact their own systems and customer interactions.

Real Time Fraud Risk Monitoring Is Now Essential

One of the biggest operational changes discussed in this episode is the need for real time fraud risk monitoring.

AI driven fraud compresses the lifecycle of fraud attacks.

What used to unfold over days can now unfold in minutes.

If fraud visibility gaps exist between teams, platforms, or systems, institutions may detect fraud only after losses occur.

Effective detection requires systems that can:

  • Correlate behavioral signals across platforms
  • Identify AI enabled social engineering patterns
  • Detect abnormal velocity shifts in activity
  • Escalate suspicious signals automatically
  • Interrupt adaptive fraud attacks while they are still active

Without unified visibility, fraud detection challenges multiply quickly.

Fraud Leadership and AI Risk Governance Must Align

Let me just assure you of something.

AI fraud threats are not just a technology problem.

They are a governance problem.

Fraud leadership and AI risk management must work together to answer key questions such as:

  • Where AI tools are embedded in customer experiences
  • How those systems are monitored for misuse
  • What controls exist to prevent AI enabled fraud
  • How quickly teams can respond when incidents occur

Fraud visibility gaps between product teams, cyber teams, and fraud teams increase risk dramatically.

Organizations that align fraud strategy with AI risk governance will be better prepared for the future of fraud prevention.

AI fraud threats are already reshaping active attack environments

AI fraud threats are not theoretical anymore.

Fraud teams are already encountering AI-driven fraud in active environments where attacks move faster, adapt more quickly, and operate across multiple platforms at the same time.

Fraudsters are no longer relying only on static phishing scripts or predictable social engineering.

AI powered scams allow attackers to iterate quickly, test defenses, and adjust tactics in real time.

That compresses the detection window dramatically.

Fraud programs that rely on slow escalation processes or retrospective analysis are going to struggle to keep up with fraud threats powered by artificial intelligence.

Agentic AI attacks are changing how fraud is executed

Agentic AI attacks are one of the biggest shifts fraud teams need to understand right now.

Unlike simple automation, autonomous fraud attacks can operate with conditional logic and adaptive behavior once they are deployed.

These systems can:

  • Test fraud detection controls across platforms
  • Modify tactics when friction appears
  • Scale attacks across multiple systems simultaneously
  • Continue operating without constant human oversight

And that creates a very different detection challenge for fraud teams.

Traditional fraud detection models built around predictable behavior struggle against adaptive fraud attacks.

Fraud prevention in the AI era requires systems that can respond just as quickly as attacks evolve.

Platform exploitation fraud expands the attack surface

AI platforms are quickly becoming part of the fraud attack surface.

When fraudsters exploit AI powered systems, platform exploitation fraud spreads through environments that users already trust.

For example, attackers may hijack AI tools to distribute malicious links, generate scam content, or amplify social engineering campaigns.

That creates layered risk for institutions.

Fraud risk in AI systems depends not only on external attackers, but also on how platforms design, deploy, and monitor their AI capabilities.

Organizations need to understand how AI misuse by fraudsters could impact their own systems and customer interactions.

Real time fraud risk monitoring is now essential

One of the biggest operational shifts discussed in this episode is the need for real time fraud risk monitoring.

AI driven fraud compresses the lifecycle of fraud attacks.

What used to unfold over days can now unfold in minutes.

If visibility gaps exist between teams, platforms, or systems, institutions may only detect fraud after losses occur.

Effective detection requires systems that can:

  • Correlate behavioral signals across platforms
  • Identify AI enabled social engineering patterns
  • Detect abnormal velocity shifts in activity
  • Escalate suspicious signals automatically
  • Interrupt adaptive fraud attacks while they are still active

Without unified visibility, fraud detection challenges multiply quickly.

Fraud leadership and AI risk governance must align

Let me just assure you of something.

AI fraud threats are not just a technology problem.

They are a governance problem.

Fraud leadership and AI risk management need to work together to answer key questions such as:

  • Where AI tools are embedded in customer experiences
  • How those systems are monitored for misuse
  • What controls exist to prevent AI enabled fraud
  • How quickly teams can respond when incidents occur

Fraud visibility gaps between product teams, cyber teams, and fraud teams increase risk dramatically.

Organizations that align fraud strategy with AI risk governance will be far better prepared for the future of fraud prevention.

Full episode transcript
Share this episode