Heterogeneous ensemble enables a universal uncertainty metric for atomistic foundation models

Journal Article (2026)
Author(s)

Kai Liu (TU Delft - Team Marcel Sluiter)

Zixiong Wei (TU Delft - Team Poulumi Dey)

Wei Gao (Texas A&M University)

Poulumi Dey (TU Delft - Team Poulumi Dey)

Marcel H.F. Sluiter (TU Delft - Team Marcel Sluiter, Universiteit Gent)

Fei Shuang (TU Delft - Team Poulumi Dey)

Research Group
Team Marcel Sluiter
DOI related publication
https://doi.org/10.1038/s41524-025-01905-x
More Info
expand_more
Publication Year
2026
Language
English
Research Group
Team Marcel Sluiter
Issue number
1
Volume number
12
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Universal machine-learning interatomic potentials (uMLIPs) are emerging as foundation models for atomistic simulation, offering near-ab initio accuracy at far lower cost. Their safe, broad deployment is limited by the absence of reliable, general uncertainty estimates. We present a unified, scalable uncertainty metric, U, built from a heterogeneous ensemble that reuses existing pretrained MLIPs. Across diverse chemistries and structures, U strongly tracks true prediction errors and robustly ranks configuration-level risk. Using U, we perform uncertainty-aware distillation to train system-specific potentials with far fewer labels: for tungsten, we match full density-functional-theory (DFT) training using 4% of the DFT data; for MoNbTaW, a dataset distilled by U supports high-accuracy potential training. By filtering numerical label noise, the distilled models can in some cases exceed the accuracy of the MLIPs trained on DFT data. This framework provides a practical reliability monitor and guides data selection and fine-tuning, enabling cost-efficient, accurate, and safer deployment of foundation models.