ΑΙhub.org
 

New AI technique sounding out audio deepfakes


by
21 November 2025



share this:

Researchers from Australia’s national science agency CSIRO, Federation University Australia and RMIT University have developed a method to improve the detection of audio deepfakes.

The new technique, Rehearsal with Auxiliary-Informed Sampling (RAIS), is designed for audio deepfake detection — a growing threat in cybercrime risks such as bypassing voice-based biometric authentication systems, impersonation and disinformation. It determines whether an audio clip is real or artificially generated (a ‘deepfake’) and maintains performance over time as attack types evolve.

In Italy earlier this year, an AI-cloned voice of its Defence Minister requested a €1M ‘ransom’ from prominent business leaders, convincing some to pay. This is just one of many examples, highlighting the need for audio deepfake detectors.

As deepfake audio technology advances rapidly, newer ‘fake techniques’ often look nothing like the older ones.

“We want these detection systems to learn the new deepfakes without having to train the model again from scratch. If you just fine-tune on the new samples, it will cause the model to forget the older deepfakes it knew before,” said joint author, Dr Kristen Moore from CSIRO’s Data61.

“RAIS solves this by automatically selecting and storing a small, but diverse set of past examples, including hidden audio traits that humans may not even notice, to help the AI learn the new deepfake styles without forgetting the old ones,” explained Dr Moore.

RAIS uses a smart selection process powered by a network that generates ‘auxiliary labels’ for each audio sample. These labels help identify a diverse and representative set of audio samples to retain and rehearse. By incorporating extra labels beyond simple ‘fake’ or ‘real’ tags, RAIS ensures a richer mix of training data, improving its ability to remember and adapt over time.

Outperforming other methods, RAIS achieves the lowest average error rate of 1.95 per cent across a sequence of five experiences. The code, available on GitHub, remains effective with a small memory buffer and is designed to maintain accuracy as attacks become more sophisticated.

“Audio deepfakes are evolving rapidly, and traditional detection methods can’t keep up,” said Falih Gozi Febrinanto, a recent PhD graduate of Federation University Australia.

“RAIS helps the model retain what it has learned and adapt to new attacks. Overall, it reduces the risk of forgetting and enhances its ability to detect deepfakes.”

“Our approach not only boosts detection performance, but also makes continual learning practical for real-world applications. By capturing the full diversity of audio signals, RAIS sets a new standard for efficiency and reliability,” said Dr Moore.

Read and download the full paper: Rehearsal with Auxiliary-Informed Sampling for Audio Deepfake Detection.




CSIRO




            AIhub is supported by:



Related posts :

Interview with Kate Larson: Talking multi-agent systems and collective decision-making

  27 Jan 2026
AIhub ambassador Liliane-Caroline Demers caught up with Kate Larson at IJCAI 2025 to find out more about her research.

#AAAI2026 social media round up: part 1

  23 Jan 2026
Find out what participants have been getting up to during the first few of days at the conference

Congratulations to the #AAAI2026 outstanding paper award winners

  22 Jan 2026
Find out who has won these prestigious awards at AAAI this year.

3 Questions: How AI could optimize the power grid

  21 Jan 2026
While the growing energy demands of AI are worrying, some techniques can also help make power grids cleaner and more efficient.

Interview with Xiang Fang: Multi-modal learning and embodied intelligence

  20 Jan 2026
In the first of our new series of interviews featuring the AAAI Doctoral Consortium participants, we hear from Xiang Fang.

An introduction to science communication at #AAAI2026

  19 Jan 2026
Find out more about our session on Wednesday 21 January.

Guarding Europe’s hidden lifelines: how AI could protect subsea infrastructure

  15 Jan 2026
EU-funded researchers are developing AI-powered surveillance tools to protect the vast network of subsea cables and pipelines that keep the continent’s energy and data flowing.


AIhub is supported by:







 













©2026.01 - Association for the Understanding of Artificial Intelligence