DistMLIP: A Distributed Inference Platform for Machine Learning Interatomic Potentials
- URL: http://arxiv.org/abs/2506.02023v1
- Date: Wed, 28 May 2025 23:23:36 GMT
- Title: DistMLIP: A Distributed Inference Platform for Machine Learning Interatomic Potentials
- Authors: Kevin Han, Bowen Deng, Amir Barati Farimani, Gerbrand Ceder,
- Abstract summary: Machine learning interatomic potentials (MLIPs) have offered a solution to scale up quantum mechanical calculations.<n>We present DistMLIP, an efficient distributed inference platform for MLIPs based on zero-redundancy, graph-level parallelization.<n>We demonstrate DistMLIP on four widely used and state-of-the-art MLIPs: CHGNet, MACE,Net, and eSEN.
- Score: 6.622327158385407
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Large-scale atomistic simulations are essential to bridge computational materials and chemistry to realistic materials and drug discovery applications. In the past few years, rapid developments of machine learning interatomic potentials (MLIPs) have offered a solution to scale up quantum mechanical calculations. Parallelizing these interatomic potentials across multiple devices poses a challenging, but promising approach to further extending simulation scales to real-world applications. In this work, we present DistMLIP, an efficient distributed inference platform for MLIPs based on zero-redundancy, graph-level parallelization. In contrast to conventional space-partitioning parallelization, DistMLIP enables efficient MLIP parallelization through graph partitioning, allowing multi-device inference on flexible MLIP model architectures like multi-layer graph neural networks. DistMLIP presents an easy-to-use, flexible, plug-in interface that enables distributed inference of pre-existing MLIPs. We demonstrate DistMLIP on four widely used and state-of-the-art MLIPs: CHGNet, MACE, TensorNet, and eSEN. We show that existing foundational potentials can perform near-million-atom calculations at the scale of a few seconds on 8 GPUs with DistMLIP.
Related papers
- chemtrain-deploy: A parallel and scalable framework for machine learning potentials in million-atom MD simulations [0.6240840318920522]
We present chemtrain-deploy, a framework that enables model-agnostic deployment of LAMMPS in MD simulations.<n>Chemtrain-deploy supports any JAX-defined semi-local potential, allowing users to exploit the functionality of LAMMPS.<n>It achieves state-of-the-art efficiency and scales to systems containing millions of atoms.
arXiv Detail & Related papers (2025-06-04T15:19:26Z) - Global Universal Scaling and Ultra-Small Parameterization in Machine Learning Interatomic Potentials with Super-Linearity [8.605514384729469]
We develop a Super-linear MLIP with both Ultra-Small parameterization and greatly expanded expressive capability, named SUS2-MLIP.<n>SUS2-MLIP outperforms the state-of-the-art MLIP models with its exceptional computational efficiency.<n>This work sheds light on incorporating physical constraints into artificial-intelligence-aided materials simulation.
arXiv Detail & Related papers (2025-02-11T06:34:31Z) - Energy & Force Regression on DFT Trajectories is Not Enough for Universal Machine Learning Interatomic Potentials [8.254607304215451]
Universal Machine Learning Interactomic Potentials (MLIPs) enable accelerated simulations for materials discovery.<n>MLIPs' inability to reliably and accurately perform large-scale molecular dynamics (MD) simulations for diverse materials.
arXiv Detail & Related papers (2025-02-05T23:04:21Z) - MAPS: Advancing Multi-Modal Reasoning in Expert-Level Physical Science [62.96434290874878]
Current Multi-Modal Large Language Models (MLLM) have shown strong capabilities in general visual reasoning tasks.<n>We develop a new framework, named Multi-Modal Scientific Reasoning with Physics Perception and Simulation (MAPS) based on an MLLM.<n>MAPS decomposes expert-level multi-modal reasoning task into physical diagram understanding via a Physical Perception Model (PPM) and reasoning with physical knowledge via a simulator.
arXiv Detail & Related papers (2025-01-18T13:54:00Z) - MLatom 3: Platform for machine learning-enhanced computational chemistry
simulations and workflows [12.337972297411003]
Machine learning (ML) is increasingly becoming a common tool in computational chemistry.
MLatom 3 is a program package designed to leverage the power of ML to enhance typical computational chemistry simulations.
The users can choose from an extensive library of methods containing pre-trained ML models and quantum mechanical approximations.
arXiv Detail & Related papers (2023-10-31T03:41:39Z) - In Situ Framework for Coupling Simulation and Machine Learning with
Application to CFD [51.04126395480625]
Recent years have seen many successful applications of machine learning (ML) to facilitate fluid dynamic computations.
As simulations grow, generating new training datasets for traditional offline learning creates I/O and storage bottlenecks.
This work offers a solution by simplifying this coupling and enabling in situ training and inference on heterogeneous clusters.
arXiv Detail & Related papers (2023-06-22T14:07:54Z) - h-analysis and data-parallel physics-informed neural networks [0.7614628596146599]
We explore the data-parallel acceleration of machine learning schemes with a focus on physics-informed neural networks (PINNs)
We detail a novel protocol based on $h$-analysis and data-parallel acceleration through the Horovod training framework.
We show that the acceleration is straightforward to implement, does not compromise training, and proves to be highly efficient and controllable.
arXiv Detail & Related papers (2023-02-17T12:15:18Z) - Partitioning Distributed Compute Jobs with Reinforcement Learning and
Graph Neural Networks [58.720142291102135]
Large-scale machine learning models are bringing advances to a broad range of fields.
Many of these models are too large to be trained on a single machine, and must be distributed across multiple devices.
We show that maximum parallelisation is sub-optimal in relation to user-critical metrics such as throughput and blocking rate.
arXiv Detail & Related papers (2023-01-31T17:41:07Z) - ETLP: Event-based Three-factor Local Plasticity for online learning with
neuromorphic hardware [105.54048699217668]
We show a competitive performance in accuracy with a clear advantage in the computational complexity for Event-Based Three-factor Local Plasticity (ETLP)
We also show that when using local plasticity, threshold adaptation in spiking neurons and a recurrent topology are necessary to learntemporal patterns with a rich temporal structure.
arXiv Detail & Related papers (2023-01-19T19:45:42Z) - Deep learning applied to computational mechanics: A comprehensive
review, state of the art, and the classics [77.34726150561087]
Recent developments in artificial neural networks, particularly deep learning (DL), are reviewed in detail.
Both hybrid and pure machine learning (ML) methods are discussed.
History and limitations of AI are recounted and discussed, with particular attention at pointing out misstatements or misconceptions of the classics.
arXiv Detail & Related papers (2022-12-18T02:03:00Z) - Einsum Networks: Fast and Scalable Learning of Tractable Probabilistic
Circuits [99.59941892183454]
We propose Einsum Networks (EiNets), a novel implementation design for PCs.
At their core, EiNets combine a large number of arithmetic operations in a single monolithic einsum-operation.
We show that the implementation of Expectation-Maximization (EM) can be simplified for PCs, by leveraging automatic differentiation.
arXiv Detail & Related papers (2020-04-13T23:09:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.