Artificial Intelligence Deepfakes Are Forging A New Path for Financial Fraud

The financial world, once a bastion of tradition, is facing a new adversary: artificial intelligence (AI). While AI has been lauded for its potential to revolutionize industries, its growing sophistication has also opened a Pandora’s box of security concerns. 

One particularly alarming development is the emergence of AI-powered services. OnlyFake, an AI image generator, is allegedly capable of generating near-undetectable counterfeit identification documents. These “deepfakes” of IDs pose a significant threat to the efficacy of current Know Your Customer (KYC) protocols, potentially allowing bad actors to bypass security measures and infiltrate financial systems.

The implications are far-reaching. Money laundering, terrorist financing, and other illicit activities could flourish in the shadows of AI-generated anonymity. In our last piece, we delved into the critical role Know Your Customer (KYC) controls play in securing the financial system from criminal activity. Today, we dissect a new threat that threatens to undermine its very effectiveness.

What is OnlyFake?

Imagine a website offering near-instant, convincing fake IDs for just $15, using technology that bypasses traditional verification systems. That’s the unsettling reality of OnlyFake, an underground website claiming to leverage “neural networks” to create realistic counterfeits.

This service disrupts the market for fake IDs, raising significant cybersecurity concerns. 404 Media verified that OnlyFake can generate highly believable IDs, potentially streamlining financial crimes like money laundering.

How Does It Work?

Forget painstaking forgery or waiting for mailed IDs. OnlyFake allows anyone to create seemingly legitimate IDs in minutes. Tests yielded a convincing California driver’s license, complete with customized details and a photo mimicking a real ID placed on a carpet, a common verification requirement for some service providers.

Using one of these generated IDs, 404 Media successfully navigated the verification process on the cryptocurrency exchange OKX.

Claims and Concerns

OnlyFake boasts “neural networks” and “generators” capable of creating 20,000 IDs daily. The owner, “John Wick,” claims the ability to generate hundreds simultaneously from Excel data. While OnlyFake claims AI involvement, experts like Hany Farid, a leading expert on digital manipulation, suspect a different technique – inserting images into ID templates. He points out inconsistencies in backgrounds that would arise with true generative AI.

The Underlying Technology and Its Implications

The core technology behind OnlyFake’s alarming efficiency is rooted in GANs and diffusion-based models. GANs involve two neural networks—the generator and the discriminator—where the generator creates fake images and the discriminator evaluates their authenticity. Over time, through continuous interaction, both networks enhance their capabilities, leading to the production of highly realistic documents. Diffusion-based models further refine this process by training on vast datasets of genuine IDs, enabling the synthesis of documents with an unprecedented level of detail and accuracy.

The New Era of Fraud

Know Your Customer (KYC) protocols are the backbone of financial security. Banks and other institutions rely on these processes to verify the identities of customers, screen for suspicious activity, and comply with regulations. They do this by carefully checking ID documents like driver’s licenses and passports to ensure they are genuine.

OnlyFake elevates automated fraud with two key features:

  • Batch creation: Enables generating numerous IDs simultaneously, perfect for fabricating identities using other AI tools.
  • Embedded generation: Creates portraits and signatures, allowing criminals to combine stolen data leaks with these synthetic identities for unprecedented realism.

These capabilities pose a significant challenge to traditional KYC (Know Your Customer) and IDV (Identity Verification) systems that rely heavily on database checks. The future of security may require a shift towards assessing the document’s authenticity itself, not just the data it holds.

AI-powered services like OnlyFake can produce forgeries that are remarkably difficult to distinguish from the real thing. This makes it increasingly easy for bad actors to slip through the cracks of KYC checks.

With a realistic fake ID in hand, a criminal can open bank accounts, apply for loans, or engage in other financial activities under a false identity. This undermines the entire system designed to keep money laundering and fraud in check.

Once fraudulent accounts are established, they become pathways for criminal activity. Money laundering – disguising the illegal source of funds – becomes easier. Worse, by concealing their identities, terrorist organizations could exploit these backdoors to fund their operations.

The utilization of AI to circumvent KYC requirements introduces a myriad of ethical and legal dilemmas. Firstly, it directly contravenes AML and KYC regulations designed to prevent financial crimes such as money laundering and terrorist financing. Moreover, the global operation of services like OnlyFake, producing documents for various countries, underscores the international legal challenges and the imperative for cross-border law enforcement cooperation.

The Bottom Line: AI-generated fake IDs directly erode the effectiveness of KYC, the foundation of financial security. They enable bad actors to bypass these safeguards and infiltrate financial systems for criminal purposes.


  • Solutions
  • Training
  • Resources
  • Support