Lea Schönherr is expert for audio file analysis. Joel Frank is specialised in AI algorithms. © RUB, Marquard

IT Security Identifying fake voice recordings

Artificial intelligence can imitate people’s voices. Scammers are already taking advantage of this on the phone. A team in Bochum is working on a solution.

Machines can use artificial intelligence to create photos or voice recordings that look or sound like those in real life. Researchers at the Horst Görtz Institute for IT Security at Ruhr-Universität Bochum are interested in how such artificially generated data, known as deepfakes, can be distinguished from real data. They found that real and fake voice recordings differ in the high frequencies. To date, deepfakes had mainly been analysed in image files. The new findings should help to recognise fake language recordings in the future.

Joel Frank from the Chair for Systems Security and Lea Schönherr from the Cognitive Signal Processing group presented their results on 7 December 2021 at the Conference on Neural Information Processing Systems, which was held online. Their research was conducted as part of the Cluster of Excellence CASA – Cybersecurity in the Age of Large-Scale Adversaries.

Large deepfake dataset generated

As a first step, Joel Frank and Lea Schönherr compiled a large dataset with around 118,000 artificially generated voice recordings. This produced about 196 hours of material in English and Japanese. “Such a dataset for audio deepfakes did not exist before,” explains Lea Schönherr. “But in order to improve the methods for detecting fake audio files, you need all this material.” To ensure that the dataset is as diverse as possible, the team used six different artificial intelligence algorithms when generating the audio snippets.

Hinweis: Beim Klick auf den Play-Button wird eine Verbindung mit einer RUB-externen Website hergestellt, die eventuell weniger strengen Datenschutzrichtlinien unterliegt und gegebenenfalls personenbezogene Daten erhebt. Weitere Informationen finden Sie in unserer Datenschutzerklärung. – Die datenschutzfreundliche Einbettung erfolgt via Embetty.

The researchers then compared the artificial audio files with recordings of real speech. They plotted the files as spectrograms showing the frequency distribution over time. The comparison revealed subtle differences in the high frequencies between real and fake files.

These spectrograms show the frequency distribution over time of a real (top) and a fake audio file (bottom). The subtle differences in the higher frequencies are marked with red circles. © RUB, Lehrstuhl für Systemsicherheit

Based on these findings, Frank and Schönherr developed algorithms that can distinguish between deepfakes and real speech. These algorithms are designed as a starting point for other researchers to develop novel detection methods.

Funding

The German Research Foundation funded the research as part of the Cluster of Excellence CASA – Cybersecurity in the Age of Large-Scale Adversaries (EXC 2092, 390781972).

Original publication

Joel Frank, Lea Schönherr: WaveFake: A data set to facilitate audio deepfake detection, Thirty-fifth Conference on Neural Information Processing Systems (NeurIPS) – Track for Datasets and Benchmarks, 2021, Online Conference, Paper-Download

Press contact

Joel Frank
Chair for Systems Security
Horst Görtz Institute for IT-Security
Ruhr-Universität Bochum
Germany
Email: joel.frank@rub.de

Lea Schönherr
Cognitive Signal Processing
Horst Görtz Institute for IT-Security
Ruhr-Universität Bochum
Germany
Phone: +49 234 32 29638
Email: lea.schoenherr@rub.de

Published

Monday
20 December 2021
9:13 am

By

Julia Weiler (jwe)

Translated by

Donata Zuber

Share