Surface Stability Modeling with Universal Machine Learning Interatomic Potentials: A Comprehensive Cleavage Energy Benchmarking Study
- URL: http://arxiv.org/abs/2508.21663v1
- Date: Fri, 29 Aug 2025 14:24:47 GMT
- Title: Surface Stability Modeling with Universal Machine Learning Interatomic Potentials: A Comprehensive Cleavage Energy Benchmarking Study
- Authors: Ardavan Mehdizadeh, Peter Schindler,
- Abstract summary: Machine learning interatomic potentials (MLIPs) have revolutionized computational materials science.<n>No systematic evaluation has assessed how well these universal MLIPs can predict cleavage energies.<n>We present a benchmark of 19 state-of-the-art uMLIPs for cleavage energy prediction.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning interatomic potentials (MLIPs) have revolutionized computational materials science by bridging the gap between quantum mechanical accuracy and classical simulation efficiency, enabling unprecedented exploration of materials properties across the periodic table. Despite their remarkable success in predicting bulk properties, no systematic evaluation has assessed how well these universal MLIPs (uMLIPs) can predict cleavage energies, a critical property governing fracture, catalysis, surface stability, and interfacial phenomena. Here, we present a comprehensive benchmark of 19 state-of-the-art uMLIPs for cleavage energy prediction using our previously established density functional theory (DFT) database of 36,718 slab structures spanning elemental, binary, and ternary metallic compounds. We evaluate diverse architectural paradigms, analyzing their performance across chemical compositions, crystal systems, thickness, and surface orientations. Our results reveal that training data composition dominates architectural sophistication: models trained on the Open Materials 2024 (OMat24) dataset, which emphasizes non-equilibrium configurations, achieve mean absolute percentage errors below 6% and correctly identify the thermodynamically most stable surface terminations in 87% of cases, without any explicit surface energy training. In contrast, architecturally identical models trained on equilibrium-only datasets show five-fold higher errors, while models trained on surface-adsorbate data fail catastrophically with a 17-fold degradation. Remarkably, simpler architectures trained on appropriate data achieve comparable accuracy to complex transformers while offering 10-100x computational speedup. These findings show that the community should focus on strategic training data generation that captures the relevant physical phenomena.
Related papers
- Equivariant Evidential Deep Learning for Interatomic Potentials [55.6997213490859]
Uncertainty quantification is critical for assessing the reliability of machine learning interatomic potentials in molecular dynamics simulations.<n>Existing UQ approaches for MLIPs are often limited by high computational cost or suboptimal performance.<n>We propose textitEquivariant Evidential Deep Learning for Interatomic Potentials ($texte2$IP), a backbone-agnostic framework that models atomic forces and their uncertainty jointly.
arXiv Detail & Related papers (2026-02-11T02:00:25Z) - Comparing the latent features of universal machine-learning interatomic potentials [1.2314765641075438]
We show that machine-learning interatomic potentials (uMLIPs) encode chemical space in significantly distinct ways.<n>We discuss how atom-level features, which are directly output by MLIPs, can be compressed into global structure-level features.
arXiv Detail & Related papers (2025-12-05T13:45:01Z) - Why Physics Still Matters: Improving Machine Learning Prediction of Material Properties with Phonon-Informed Datasets [0.32622301272834514]
We assess the effectiveness of graph neural network (GNN) models trained on two fundamentally different types of datasets.<n>As a case study, we address the challenging task of predicting electronic and mechanical properties of a prototypical family of optoelectronic materials.<n>We find that the phonons-informed model consistently outperforms the randomly trained counterpart, despite relying on fewer data points.
arXiv Detail & Related papers (2025-11-19T08:16:10Z) - Foundation Models for Discovery and Exploration in Chemical Space [57.97784111110166]
MIST is a family of molecular foundation models trained on large unlabeled datasets.<n>We demonstrate the ability of these models to solve real-world problems across chemical space.
arXiv Detail & Related papers (2025-10-20T17:56:01Z) - From Physics to Machine Learning and Back: Part II - Learning and Observational Bias in PHM [52.64097278841485]
Review examines how incorporating learning and observational biases through physics-informed modeling and data strategies can guide models toward physically consistent and reliable predictions.<n>Fast adaptation methods including meta-learning and few-shot learning are reviewed alongside domain generalization techniques.
arXiv Detail & Related papers (2025-09-25T14:15:43Z) - BLIPs: Bayesian Learned Interatomic Potentials [47.73617239750485]
Machine Learning Interatomic Potentials (MLIPs) are becoming a central tool in simulation-based chemistry.<n>MLIPs do not provide uncertainty estimates by construction, which are fundamental to guide active learning pipelines.<n>BLIP is a scalable, architecture-agnostic variational Bayesian framework for training or fine-tuning MLIPs.
arXiv Detail & Related papers (2025-08-19T17:28:14Z) - Performance of universal machine-learned potentials with explicit long-range interactions in biomolecular simulations [21.340102594388348]
Universal machine-learned potentials promise transferable accuracy across compositional and vibrational degrees of freedom.<n>This work systematically evaluates equivariant message-passing architectures trained on the SPICE-v2 dataset with and without explicit long-range dispersion and electrostatics.
arXiv Detail & Related papers (2025-08-14T17:08:34Z) - Evaluating Universal Machine Learning Force Fields Against Experimental Measurements [15.863801293927635]
Universal machine learning force fields (UMLFFs) promise to revolutionize materials science by enabling rapid atomistic simulations across the periodic table.<n>Here, we present UniFFBench, a comprehensive framework for evaluating experimental bondingFFs against experimental measurements of 1,500 carefully curated mineral structures.<n>Our systematic evaluation of six state-of-the-artFFs reveals a substantial reality gap: models achieving impressive performance on computational benchmarks often fail when confronted with experimental complexity.
arXiv Detail & Related papers (2025-08-07T18:21:39Z) - Iterative Pretraining Framework for Interatomic Potentials [46.53683458224917]
We propose Iterative Pretraining for Interatomic Potentials (IPIP) to improve predictive performance of MLIP models.<n>IPIP incorporates a forgetting mechanism to prevent iterative training from converging to suboptimal local minima.<n>Compared to general-purpose force fields, this approach achieves over 80% reduction in prediction error and up to 4x speedup in the challenging Mo-S-O system.
arXiv Detail & Related papers (2025-07-27T03:59:41Z) - Spatiotemporal Field Generation Based on Hybrid Mamba-Transformer with Physics-informed Fine-tuning [46.67399627400437]
This research confronts the challenge of substantial physical equation discrepancies in the generation of physical fields through trained models.<n>A physical field generation model, named HMT-PF, is developed based on the hybrid Mamba-Transformer architecture.
arXiv Detail & Related papers (2025-05-16T14:40:56Z) - Accurate, transferable, and verifiable machine-learned interatomic potentials for layered materials [0.0]
Twisted layered van-der-Waals materials often exhibit unique electronic and optical properties absent in their non-twisted counterparts.<n>Here, we introduce a split machine-learned interatomic potential and dataset curation approach that separates intralayer and interlayer interactions.<n>Our approach integrates seamlessly with various intralayer and interlayer interaction models, enabling computationally tractable relaxation of moir'e materials.
arXiv Detail & Related papers (2025-03-19T17:14:02Z) - Overcoming systematic softening in universal machine learning interatomic potentials by fine-tuning [3.321322648845526]
Machine learning interatomic potentials (MLIPs) have introduced a new paradigm for atomic simulations.
Recent advancements have seen the emergence of universal MLIPs (uMLIPs) that are pre-trained on diverse materials datasets.
We show that their performance in extrapolating to out-of-distribution complex atomic environments remains unclear.
arXiv Detail & Related papers (2024-05-11T22:30:47Z) - Accurate machine learning force fields via experimental and simulation
data fusion [0.0]
Machine Learning (ML)-based force fields are attracting ever-increasing interest due to their capacity to span scales of classical interatomic potentials at quantum-level accuracy.
Here we leverage both Density Functional Theory (DFT) calculations and experimentally measured mechanical properties and lattice parameters to train an ML potential of titanium.
We demonstrate that the fused data learning strategy can concurrently satisfy all target objectives, thus resulting in a molecular model of higher accuracy compared to the models trained with a single source data.
arXiv Detail & Related papers (2023-08-17T18:22:19Z) - BIGDML: Towards Exact Machine Learning Force Fields for Materials [55.944221055171276]
Machine-learning force fields (MLFF) should be accurate, computationally and data efficient, and applicable to molecules, materials, and interfaces thereof.
Here, we introduce the Bravais-Inspired Gradient-Domain Machine Learning approach and demonstrate its ability to construct reliable force fields using a training set with just 10-200 atoms.
arXiv Detail & Related papers (2021-06-08T10:14:57Z) - OrbNet: Deep Learning for Quantum Chemistry Using Symmetry-Adapted
Atomic-Orbital Features [42.96944345045462]
textscOrbNet is shown to outperform existing methods in terms of learning efficiency and transferability.
For applications to datasets of drug-like molecules, textscOrbNet predicts energies within chemical accuracy of DFT at a computational cost that is thousand-fold or more reduced.
arXiv Detail & Related papers (2020-07-15T22:38:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.