ΑΙhub.org
 

FLeet: putting machine learning in your pocket


by
02 June 2021



share this:
mobile phone in bag

By Tanya Petersen

EPFL/INRIA research shows that it is possible for our mobile devices to conduct machine learning as part of a distributed network, without giving big global tech companies access to our data.

Every time we read news online or search for somewhere to eat out, big tech collects huge amounts of our behavioral data. Google and Facebook, for example, say they do this to improve service and make our online experiences more personalized.

Their artificial intelligence (AI) filters news for articles they think you will be more interested in, offers videos similar to ones you have watched before, or helps you to find a restaurant based on ones you have previously liked. On the flipside, this data is also used to target advertising towards you, and can be shared with third parties, a key reason that digital privacy issues are so prominent.

Now, new research from the Distributed Computing Laboratory and Scalable Computing Systems Laboratory, part of EPFL’s School of Computer and Communication Science (IC) and the French National Institute for Research in Digital Science and Technology (INRIA) has shown that it is possible for machine learning to be undertaken on our mobile devices in real time, without compromising their functionality, and without having to share our data.

Conducted in the context of the EPFL/INRIA joint laboratory, the work introduces FLeet, a revolution in what is known as Federated Learning – a global model trained with updates computed on mobile devices while keeping the data of users local. Federated Learning is very appealing for its privacy benefits but because it is designed to have no energy or performance impact on mobile devices it’s not suitable for applications that require frequent updates, such as news recommendations.

FLeet combines the privacy of standard Federated Learning with the precision of online learning thanks to two core components: I-Prof, a new lightweight profiler that predicts and controls the impact of learning tasks on mobile devices, and AdaSGD, a new adaptive learning algorithm that is resilient to delayed updates.

As one of the paper’s authors, EPFL Professor Rachid Guerraoui explains, it turns out that today our smartphones have both the data and battery power to enable distributed machine learning to take place.

“With FLeet it is possible, while you are using your mobile phone, to use some of its spare power towards machine learning tasks without having to worry that your call or internet search will be interrupted. One of the reasons this is important, and we don’t want the machine learning to only be happening when you are asleep and your phone is on charge, is because sometimes we want, and need, real time information.”

Professor Anne-Marie Kermarrec was also an author on the paper, “What we have shown is that if we put all our phones together they start constituting big computing power to match the likes of Google and that gives people alternatives to relying on centralized, powerful computer farms. This is truly collaborative learning where local models are aggregated and contribute to the global model but you don’t share raw data and that protects your privacy, a huge motivation behind developing this kind of architecture.”

For now, FLeet is a prototype that shows what is possible. Guerraoui and Kermarrec say the next step is to continue work on developing a usable end product as well as research into other aspects of FLeet’s design, including how to make the system secure against possible attacks.

“There is a huge trend nowadays to try to bring machine learning activities back to the user because, after all, the data comes from us and we should be able to decide what happens on our own devices and with our own data. Models like FLeet will give people a choice by providing an alternative to the big tech players if they want it,” concluded Kermarrec.

FLeet: Online Federated Learning via Staleness Awareness and Performance Prediction, authored by Georgios Damaskinos (Facebook) Rachid Guerraoui, Anne-Marie Kermarrec (EPFL), Rhicheek Patra (Oracle Labs), Francois Taiani (INRIA/University of Rennes) and Vlad Nitu (CNRS) recently won best conference paper at the 2020 ACM/IFIP Middleware Conference, a major forum for the discussion of innovations and recent scientific advances of middleware systems with a focus on the design, implementation, deployment, and evaluation of distributed systems, platforms and architectures for computing, storage, and communication.

AIhub focus issue on reduced inequalities

tags: ,


EPFL

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

A multi-armed robot for assisting with agricultural tasks

and   27 Mar 2026
How can a robot safely manipulate branches to reveal hidden flowers while remaining aware of interaction forces and minimizing damage?

Resource-constrained image generation and visual understanding: an interview with Aniket Roy

  26 Mar 2026
Aniket tells us about his research exploring how modern generative models can be adapted to operate efficiently while maintaining strong performance.

RWDS Big Questions: how do we highlight the role of statistics in AI?

  25 Mar 2026
Next in our series, the panel explores the statistical underpinning of AI.

A history of RoboCup with Manuela Veloso

  24 Mar 2026
Find out how RoboCup got started and how the competition has evolved, from one of the co-founders.

Information-driven design of imaging systems

  23 Mar 2026
Framework that enables direct evaluation and optimization of imaging systems based on their information content.

Machine learning framework to predict global imperilment status of freshwater fish

  20 Mar 2026
“With our model, decision makers can deploy resources in advance before a species becomes imperiled.”

Interview with AAAI Fellow Yan Liu: machine learning for time series

  19 Mar 2026
Hear from 2026 AAAI Fellow Yan Liu about her research into time series, the associated applications, and the promise of physics-informed models.

A principled approach for data bias mitigation

  18 Mar 2026
Find out more about work presented at AIES 2025 which proposes a new way to measure data bias, along with a mitigation algorithm with mathematical guarantees.



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence