Humans Unable to Reliably Detect Deepfake Speech


Humans cannot detect deepfake speech 27% the time, researchers from University College London (UCL) have found during a recent study.

The study presented 529 individuals with genuine and deepfake audio samples and asked them to identify the deepfakes. Participants could only identify the fake audio 73% of the time, although detection accuracy improved by 3.84% on average after they received training to recognize aspects of deepfake speech.

The researchers used a text-to-speech (TTS) algorithm trained on two publicly available datasets to produce the deepfake speech samples. These were run in English and Mandarin to understand if language can affect detection performance and decision-making rationale.

First author of the study, Kimberly Mai, commented: “Our findings confirm that humans are unable to reliably detect deepfake speech, whether or not they have received training to help them spot artificial content. It’s also worth noting that the samples that we used in this study were created with algorithms that are relatively old, which raises the question whether humans would be less able to detect deepfake speech created using the most sophisticated technology available now and in the future.”

The study is the first to assess human ability to detect artificially generated speech in a language other than English. The researchers are now planning to develop better automated speech detectors as part of efforts to create detection capabilities for deepfake audio and imagery.

Why Deepfakes Are a Growing Danger

Deepfake technology – synthetic media generated by artificial intelligence and machine learning algorithms – has become increasingly realistic in recent years. It encompasses audio, video and image manipulations or completely fake creations content of individuals, often public figures.

In 2019, the CEO of a UK-based energy company was duped into transferring $243,000 to fraudsters after receiving a phone call from someone who claimed to be the firm’s chief executive. In fact, AI voice technology was used to spoof the German chief executive’s voice. This was the first recorded case of a deepfake scam.

In another case, in October 2021, court documents revealed that a Hong Kong bank had been swindled out of $35m following an elaborate deepfake plot. The fraudsters used ‘deep voice’ technology to clone the voice of a company director about to make an acquisition and asked the bank to authorize transfers worth $35m.

In July 2023, UK-based money saving expert Martin Lewis expressed shock at a deepfake video that impersonated him promoting an investment scam.

There are increasing calls to regulate deepfakes due to its potential to be used for nefarious activities, such as fraud and spreading misinformation. For example, the UK’s Online Safety Bill includes provisions to criminalize the sharing of ‘deepfake porn.’



Source link