ΑΙhub.org
 

Machine learning for atomic-scale simulations: balancing speed and physical laws


by , and
10 October 2025



share this:

Taken from simulation of a nitrogen molecule on an iron surface exploding with non-conservative forces. See below for the full simulation.

When we want to understand how matter behaves, the real action happens at the atomic scale. Heating of water, a chemical reaction in a battery, the way proteins fold in our cells, or how a catalyst works to convert carbon dioxide into useful fuels, all of these processes are governed by the motions and interactions of atoms.

Atomic-scale simulations give us a way to explore the microscopic behavior of matter, by tracking how atoms move under the laws of quantum mechanics. These simulations have become essential across physics, chemistry, biology, and materials science. They test hypotheses that experiments cannot easily probe and help design new materials before they are synthesized and tested in the lab.

The catch is that accuracy comes at a huge computational cost. Simulating even a few hundred atoms with quantum mechanical calculations can be so expensive that the simulation covers only millionths of a second of real time. That’s not enough to see most interesting processes unfold. To capture chemical reactions, protein dynamics, or long-term materials stability, we would need simulations that run thousands or even millions of times longer.

The role of machine learning

Machine learning (ML) has transformed this picture. Instead of solving the equations of quantum mechanics at every simulation step, we can train ML models to mimic them. These machine-learned interatomic potentials (MLIPs) learn the relationship between the arrangement of atoms and the forces they exert on each other, which ultimately drive their dynamics.

With MLIPs, simulations that once took months on a supercomputer can now be run in days or even hours, often with comparable accuracy. This acceleration has made it possible to explore larger systems and longer timescales than were ever practical with first-principles calculations. But, as is often the case in ML, there is a tension between speed and fidelity. How much of the underlying physics can we safely “shortcut” without breaking the simulation? At the heart of this tension lies the question of whether machine learning models should faithfully enforce physical laws, or whether approximations that break them might be acceptable if the result is faster or more accurate predictions.

The “dark side” of non-conservative forces

One shortcut that has gained popularity is to predict atomic forces directly, rather than computing them as derivatives of an energy with respect to the atomic positions. This avoids a computationally expensive differentiation step and makes models faster to train and run. However, forces computed as derivatives automatically conserve energy, while directly predicted ones do not. In physics, such forces are called non-conservative. And if energy conservation is broken, the entire simulation can fail catastrophically.

In the paper The dark side of the forces: assessing non-conservative force models for atomistic machine learning, we investigate what happens when these non-conservative forces are used in practice and propose practical and efficient solutions to fix the resulting problems. We find that simulations driven by non-conservative forces can quickly become unstable. Geometry optimizations, which are used to find the most stable atomic structures, may fail to converge. Molecular dynamics runs — meant to simulate motion of atoms — can exhibit runaway heating, with energy drifting at rates that correspond to billions of degrees per second. Clearly, no real physical system behaves this way and this makes purely non-conservative models unreliable for production use.

However, we also identify a promising solution: hybrid models. By pre-training models on direct forces to gain efficiency, and then fine-tuning them with conservative forces, it is possible to recover stability while still enjoying almost all the computational speed-up. Similarly, when using the model to perform simulations, most evaluations can be made with the fast direct forces, using conservative forces only rarely as a correction. In other words, non-conservative forces are not useless — they just need to be combined carefully with physically grounded methods to avoid their “dark side”.

A conservative simulation of a nitrogen molecule on an iron surface (time to run: 36 min).

The same simulation, exploding with non-conservative forces (time to run: 17.3 min).

The same simulation, combining the two types of forces with a multiple-time-stepping algorithm (time to run: 21.7 min).

Lessons and outlook

This work highlights the opportunities and challenges of optimizing the speed of machine learning for atomic-scale simulations. On one hand, shortcuts that ignore physics can lead to spectacular failures — unstable trajectories, unphysical heating, unreliable predictions. On the other hand, only ML approaches that are physics-aware (up to some point) are the only ones which can provide physically-correct simulations. The likely path forward is not to entirely sacrifice physics with ML, but to combine the two. Hybrid approaches that merge machine learning efficiency with physical constraints can provide the best of both worlds.

Looking ahead, there are also further opportunities to rethink the very framework of molecular dynamics with machine learning. In a follow-up work, FlashMD: long-stride, universal prediction of molecular dynamics, we explore how ML can be used not just to accelerate force calculations, but to directly predict atomic trajectories over much longer time steps. This approach allows simulations to reach timescales that are otherwise completely out of reach, while still incorporating mechanisms to enforce energy conservation and preserve qualitative physical behavior.

As these methods mature, researchers will be able to simulate larger systems, over longer timescales, and at higher accuracy than ever before. This will accelerate discoveries in energy storage, drug design, catalysis, and countless other areas where atomic-scale insight is key. Machine learning for atomic simulations is not just about speed — it’s about finding the right balance between efficiency and physical truth. By staying grounded in the laws of nature while embracing the flexibility of ML, we can move closer to solving pressing scientific and technological challenges.



tags: ,


Filippo Bigi is a PhD student at EPFL.
Filippo Bigi is a PhD student at EPFL.

Marcel Langer is a postdoctoral researcher in the Laboratory of Computational Science and Modeling (COSMO) at EPFL.
Marcel Langer is a postdoctoral researcher in the Laboratory of Computational Science and Modeling (COSMO) at EPFL.

Michele Ceriotti is a professor at EPFL.
Michele Ceriotti is a professor at EPFL.

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

monthly digest

AIhub monthly digest: March 2026 – time series, multiplicity, and the history of RoboCup

  31 Mar 2026
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

What I’ve learned from 25 years of automated science, and what the future holds: an interview with Ross King

  30 Mar 2026
We launch our new series with a conversation with Ross King - a pioneer in the field of AI-enabled scientific discovery.

A multi-armed robot for assisting with agricultural tasks

and   27 Mar 2026
How can a robot safely manipulate branches to reveal hidden flowers while remaining aware of interaction forces and minimizing damage?

Resource-constrained image generation and visual understanding: an interview with Aniket Roy

  26 Mar 2026
Aniket tells us about his research exploring how modern generative models can be adapted to operate efficiently while maintaining strong performance.

RWDS Big Questions: how do we highlight the role of statistics in AI?

  25 Mar 2026
Next in our series, the panel explores the statistical underpinning of AI.

A history of RoboCup with Manuela Veloso

  24 Mar 2026
Find out how RoboCup got started and how the competition has evolved, from one of the co-founders.

Information-driven design of imaging systems

  23 Mar 2026
Framework that enables direct evaluation and optimization of imaging systems based on their information content.

Machine learning framework to predict global imperilment status of freshwater fish

  20 Mar 2026
“With our model, decision makers can deploy resources in advance before a species becomes imperiled.”



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence