DEV Community

Thomas
Thomas

Posted on

Realtime deepfake software is a SaaS product now

I've been half-following the deepfake-in-the-wild beat for a while. Most of it has been static image stuff fake profile photos, AI generated headshots on LinkedIn, that kind of thing. I run suspicious images through AI or Not when something looks off, flag it, move on.

But the 404 Media investigation into "HELLO BOSS" software shifted my sense of where the floor actually is. This isn't someone uploading a faked image. This is a live video call where the person on screen is not the person on screen.

What the pipeline actually looks like

The software they describe isn't magic it's a real-time face swap layer that sits between the camera input and whatever video call software the scammer is using. The rough architecture:

[Scammer's webcam]
        ↓
[Face detection + landmark extraction]
        ↓
[Target face model (pre-trained on victim's photos)]
        ↓
[Rendered output frame]
        ↓
[Virtual camera driver — OBS, v4l2loopback, etc.]
        ↓
[Zoom / WhatsApp / Teams / any WebRTC app]
Enter fullscreen mode Exit fullscreen mode

The virtual camera driver is the key piece most people miss. Tools like OBS Virtual Camera on Windows/Mac or v4l2loopback on Linux let you present any video source as a system webcam. The calling app has no idea it's not getting real hardware input.

The face-swap model itself runs inference on every frame typically 24–30 fps which used to require a beefy GPU. Consumer grade hardware can handle it now, and cloud GPU instances are cheap enough that you can rent the compute if you don't own it.

"Hello boss" isn't a technical term it's a script

The name comes from one of the primary use cases: impersonating a company executive on a video call to authorize a wire transfer. Subordinate gets a call from what looks like their CEO on screen, hears a voice that's been cloned or at least pitch-shifted, and gets told to move money somewhere.

The "hello boss" phrasing is the greeting on the other end the scammer picking up a call from someone who thinks they're reaching their boss, not the scammer impersonating the boss outbound. Either way, the social engineering depends entirely on the live video being convincing enough to short-circuit skepticism.

This same stack powers romance scams and "pig butchering" investment fraud, where the point is sustained trust over weeks, not a one time wire transfer. A static fake photo stops working the moment someone asks for a video call. A real time face swap keeps the fiction going.

Why this breaks video KYC assumptions

A lot of identity verification flows have converged on "liveness check + face match" as the standard:

1. User holds up ID document → OCR extracts name, DOB, document number
2. User records short selfie video → liveness detection (blink, turn head)
3. Face on selfie matches face on document → pass
Enter fullscreen mode Exit fullscreen mode

This pipeline assumes the face in the selfie video is the person's real face. Real time face swap defeats step 3 entirely if the attacker pre trains their model on photos of the person whose identity they're stealing. It also defeats liveness checks the swap handles arbitrary head movements and expressions in real time, so asking someone to blink or smile doesn't help.

Some vendors are adding texture analysis, illumination consistency checks, and temporal coherence scoring to catch the artifacts that face swap models still produce at frame boundaries and occlusion edges. But that arms race is already underway and the defenders are not obviously winning.

What I'd actually do differently if I were building this today

  1. Don't trust video alone. Pair any video verification step with a second channel SMS OTP, authenticator app, document scan from a different session so compromising the video doesn't compromise the whole flow.

  2. Log the raw video stream for async review. Real-time detectors aren't reliable enough to be gatekeepers. Use them as signals, not hard blocks, and let a human or a more thorough model review borderline cases after the fact.

  3. Add device fingerprinting. Face swap pipelines route through a virtual camera driver. The camera device name exposed by browser WebRTC APIs (MediaDeviceInfo.label) will often be "OBS Virtual Camera" or similar. That's not a perfect signal, but it's a cheap one worth logging.

  4. Test your liveness checks against an actual face swap. There are open source models you can run locally. If your liveness check passes, you need to know now, not when a fraud team calls you.

  5. Assume the video is synthetic and design accordingly. Treat video verification as a corroborating signal, not a root-of-trust.

Checking clips yourself

When I see video circulating that seems suspicious a celebrity endorsing something, an executive making a statement AI or Not is what I pull up first. It handles video files, not just images, so you can actually run the clip rather than screenshotting a frame and hoping the compression didn't wash out the artifacts. It's not a forensic lab, but it's fast and the confidence scores are useful for triage.

The problem the 404 Media story describes is harder to catch after the fact because there usually isn't a recording it happened on a live call. But for anything that did get recorded, that kind of tooling matters.

The SaaS part is the part that scales

What makes this story different from "deepfakes exist, film at 11" is the distribution model. The software described in the 404 Media piece is sold through Telegram channels at subscription prices. That means:

  • Low barrier to entry. You don't need to train a model or write code.
  • Sellers have support channels, update cadences, and refund policies.
  • Supply scales with demand, not with technical skill.

This is the same trajectory malware took. Ransomware-as-a-service normalized the idea that you could be a criminal without being a programmer. Deepfake-as-a-service is doing the same thing for identity fraud.

The underlying models will keep improving. The virtual camera trick is already table stakes. At some point, real-time voice cloning (already fairly mature) and real-time video swap running together on consumer hardware will be seamless enough that "get on a video call" is no longer a meaningful trust signal.

That's not a prediction it's a present tense engineering problem.

Top comments (0)