ΑΙhub.org
monthly digest
 

AIhub monthly digest: September 2024 – real-time payments, evaluating dataset diversity, and AfriClimate AI at the Deep Learning Indaba


by
26 September 2024



share this:
Panda and tiger reading

Welcome to our monthly digest, where you can catch up with any AIhub stories you may have missed, peruse the latest news, recap recent events, and more. This month, we learn about a framework to evaluate diversity in datasets, find out how banks may strategically mitigate their risk from fraud in real-time payment systems, and hear about the AfriClimate AI workshop at the Deep Learning Indaba.

Interview with Jerone Andrews: a framework towards evaluating diversity in datasets

In their paper Measure Dataset Diversity. Don’t Just Claim It, Jerone Andrews and colleagues propose using measurement theory from the social sciences as a framework to improve the collection and evaluation of diverse machine learning datasets. We spoke to Jerone about this work, which won a best paper award at ICML 2024.

Interview with Katherine Mayo: An agent-based analysis of real-time payments and fraud risk mitigation

Real-time payments offer a fast processing time (of around 10 seconds), allowing for near-immediate receipt of funds. However, these systems are a target for fraud. In this interview, Katherine Mayo tells us about an agent-based analysis of real-time payments, and what this reveals in terms of bank strategies for mitigating fraud.

IJCAI doctoral consortium

The International Joint Conference on Artificial Intelligence (IJCAI) doctoral consortium provides a group of PhD students the opportunity to present their work and hear from established researchers in the field. The chairs of this year’s consortium, Anita Raja and Jihie Kim, pick out some of their highlights in this summary.

AfriClimate AI at the Deep Learning Indaba

AfriClimate AI is a grassroots community dedicated to harnessing the power of AI for a sustainable, prosperous and climate-resilient Africa. Conversations at the Deep Learning Indaba in 2023 sparked the formation of AfriClimate AI, and the team behind this initiative returned to the Indaba this year to hold a workshop. You can read a summary of this event here.

Generative AI resources list updated

Last year, we started a collection of articles, opinion pieces, videos and resources relating to large language models, and other generative models. We periodically update this list to include the latest resources, and we’re now on the fifth iteration. Check it out here.

All Questions Answered: AI for citizens

The latest in the series of CLAIRE (Confederation of Laboratories for Artificial Intelligence Research in Europe) All Questions Answered (AQuA) sessions continued with the panel tackling the topic of “AI for citizens”. You can find the recording of the event here.

Global AI Research Agenda and Playbook

This week saw the release of two complimentary reports from the US Government: the Global AI Research Agenda and the AI in Global Development Playbook. These documents are designed to guide future research on artificial intelligence and its use in advancing the UN Sustainable Development Goals.

Artifice and Intelligence

In an article titled Artifice and Intelligence, Emily Tucker explains why the Center on Privacy & Technology at Georgetown Law has stopped using the terms “artificial intelligence,” “AI,” and “machine learning” in their institutional vocabulary. Instead, they will aim to more accurately describe each individual technology within the context of the specific case that it arises.

Predicting: The future of health?

In a recent report from the Ada Lovelace Institute, Maili Raven-Adams and Andrew Strait assess the potential, risks and appropriate role of AI-powered genomic health prediction in the UK health system.

Building a LLM from scratch

In this YouTube video, Sebastian Raschka gives a three-hour coding workshop which takes viewers through the process of building a large language model. This tutorial is aimed at coders interested in understanding the building blocks of LLMs, how they work, and how to code them from the ground up in PyTorch.


Our resources page
Our events page
Seminars in 2024
AAAI/ACM SIGAI Doctoral Consortium interview series
AAAI Fellows 2024 interview series
AI around the world focus series
New voices in AI series



tags:


Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

A multi-armed robot for assisting with agricultural tasks

and   27 Mar 2026
How can a robot safely manipulate branches to reveal hidden flowers while remaining aware of interaction forces and minimizing damage?

Resource-constrained image generation and visual understanding: an interview with Aniket Roy

  26 Mar 2026
Aniket tells us about his research exploring how modern generative models can be adapted to operate efficiently while maintaining strong performance.

RWDS Big Questions: how do we highlight the role of statistics in AI?

  25 Mar 2026
Next in our series, the panel explores the statistical underpinning of AI.

A history of RoboCup with Manuela Veloso

  24 Mar 2026
Find out how RoboCup got started and how the competition has evolved, from one of the co-founders.

Information-driven design of imaging systems

  23 Mar 2026
Framework that enables direct evaluation and optimization of imaging systems based on their information content.

Machine learning framework to predict global imperilment status of freshwater fish

  20 Mar 2026
“With our model, decision makers can deploy resources in advance before a species becomes imperiled.”

Interview with AAAI Fellow Yan Liu: machine learning for time series

  19 Mar 2026
Hear from 2026 AAAI Fellow Yan Liu about her research into time series, the associated applications, and the promise of physics-informed models.

A principled approach for data bias mitigation

  18 Mar 2026
Find out more about work presented at AIES 2025 which proposes a new way to measure data bias, along with a mitigation algorithm with mathematical guarantees.



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence