Subscribe
By subscribing you agree to with our Privacy Policy and provide consent to receive updates from our company.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

The FBI are cracking down on people using deepfakes to apply for remote jobs

Companies looking to hire somebody with IT abilities might want to ask themselves whether the individual they are interviewing is real or not.

The FBI has uncovered numerous cases involving serious technology crimes, making it essential for companies to remain vigilant against impostors. These fraudsters use images, videos, and voice recordings to secure positions in technology, programming, database, and software firms. By exploiting stolen data, they can convincingly impersonate anyone they choose.

Many of these fake job applicants may have gained access to sensitive consumer or corporate information, including financial and proprietary data, suggesting that their intentions extend beyond simple fraud. They may aim to steal valuable information while deceiving the company. The extent of successful fake job applications remains unclear compared to those that have been detected and reported.

A more alarming possibility is that some of these impostors may have accepted job offers, received salaries, and later been arrested. Instances have been reported where voice spoofing techniques were used during online interviews, with the candidate’s lip movements not matching the audio. In some cases, applicants coughed or sneezed, yet the video spoofing software failed to detect these discrepancies.

In May, the FBI issued a warning to businesses about North Korean government operatives seeking remote IT and other technical jobs. These impostors often use fake documents and credentials to secure remote work through platforms like Upwork and Fiverr. The federal agency's report detailed how some fraudsters used multiple shell companies to conceal their identities, making detection even more challenging.

While deepfake technology has advanced significantly, some of the more rudimentary attempts still result in mismatches between fake voices and the speaker’s mouth movements. Detecting these fake videos can be difficult, especially if one is not actively searching for them. Creating a lifelike human in a video is more complex than it might appear, and unnoticed fakes can slip through the cracks.

Researchers at Carnegie Mellon University have developed artificial intelligence capable of recognising edited videos with an accuracy ranging from 30% to 97%. Those familiar with identifying certain visual inconsistencies, such as incorrect shadows or unnatural skin textures, can detect phoney videos more effectively.

September 5, 2022
Super Bowl 2026, A Game to Remember
February 10, 2026

Catch all the highlights from Super Bowl 2026, including Bad Bunny’s halftime show, Seattle Seahawks’ big win, and the most memorable commercials that got everyone talking.

Read more
What Synthesia’s $4bn Valuation Says About Global Hiring in Emerging Tech
January 28, 2026

What Synthesia’s $4bn valuation reveals about global hiring trends across AI, Web3 and emerging tech, and why competition for senior talent is intensifying worldwide.

Read more
Web3 & Crypto Salary Benchmarks, What Top Engineering Talent Really Costs
January 6, 2026

Real salary benchmarks for Solidity, Rust and ZK engineers in Web3.

Read more