ΑΙhub.org
 

New AI technique sounding out audio deepfakes


by
21 November 2025



share this:

Researchers from Australia’s national science agency CSIRO, Federation University Australia and RMIT University have developed a method to improve the detection of audio deepfakes.

The new technique, Rehearsal with Auxiliary-Informed Sampling (RAIS), is designed for audio deepfake detection — a growing threat in cybercrime risks such as bypassing voice-based biometric authentication systems, impersonation and disinformation. It determines whether an audio clip is real or artificially generated (a ‘deepfake’) and maintains performance over time as attack types evolve.

In Italy earlier this year, an AI-cloned voice of its Defence Minister requested a €1M ‘ransom’ from prominent business leaders, convincing some to pay. This is just one of many examples, highlighting the need for audio deepfake detectors.

As deepfake audio technology advances rapidly, newer ‘fake techniques’ often look nothing like the older ones.

“We want these detection systems to learn the new deepfakes without having to train the model again from scratch. If you just fine-tune on the new samples, it will cause the model to forget the older deepfakes it knew before,” said joint author, Dr Kristen Moore from CSIRO’s Data61.

“RAIS solves this by automatically selecting and storing a small, but diverse set of past examples, including hidden audio traits that humans may not even notice, to help the AI learn the new deepfake styles without forgetting the old ones,” explained Dr Moore.

RAIS uses a smart selection process powered by a network that generates ‘auxiliary labels’ for each audio sample. These labels help identify a diverse and representative set of audio samples to retain and rehearse. By incorporating extra labels beyond simple ‘fake’ or ‘real’ tags, RAIS ensures a richer mix of training data, improving its ability to remember and adapt over time.

Outperforming other methods, RAIS achieves the lowest average error rate of 1.95 per cent across a sequence of five experiences. The code, available on GitHub, remains effective with a small memory buffer and is designed to maintain accuracy as attacks become more sophisticated.

“Audio deepfakes are evolving rapidly, and traditional detection methods can’t keep up,” said Falih Gozi Febrinanto, a recent PhD graduate of Federation University Australia.

“RAIS helps the model retain what it has learned and adapt to new attacks. Overall, it reduces the risk of forgetting and enhances its ability to detect deepfakes.”

“Our approach not only boosts detection performance, but also makes continual learning practical for real-world applications. By capturing the full diversity of audio signals, RAIS sets a new standard for efficiency and reliability,” said Dr Moore.

Read and download the full paper: Rehearsal with Auxiliary-Informed Sampling for Audio Deepfake Detection.




CSIRO




            AIhub is supported by:



Related posts :



Learning robust controllers that work across many partially observable environments

  20 Nov 2025
Exploring designing controllers that perform reliably even when the environment may not be precisely known.

ACM SIGAI Autonomous Agents Award 2026 open for nominations

  19 Nov 2025
Nominations are solicited for the 2026 ACM SIGAI Autonomous Agents Research Award.

Interview with Mario Mirabile: trust in multi-agent systems

  18 Nov 2025
We meet ECAI Doctoral Consortium participant, Mario, to find out more about his research.

Review of “Exploring metaphors of AI: visualisations, narratives and perception”

and   17 Nov 2025
A curated research session at the Hype Studies Conference, “(Don’t) Believe the Hype?!” 10-12 September 2025, Barcelona.

Designing value-aligned autonomous vehicles: from moral dilemmas to conflict-sensitive design

  13 Nov 2025
Autonomous systems increasingly face value-laden choices. This blog post introduces the idea of designing “conflict-sensitive” autonomous traffic agents that explicitly recognise, reason about, and act upon competing ethical, legal, and social values.

Learning from failure to tackle extremely hard problems

  12 Nov 2025
This blog post is based on the work "BaNEL: Exploration posteriors for generative modeling using only negative rewards".

How AI can improve storm surge forecasts to help save lives

  10 Nov 2025
Looking at how AI models can help provide more detailed forecasts more quickly.



 

AIhub is supported by:






 












©2025.05 - Association for the Understanding of Artificial Intelligence