Investigating the Behavior of Diffusion Models for Accelerating
Electronic Structure Calculations
- URL: http://arxiv.org/abs/2311.01491v1
- Date: Thu, 2 Nov 2023 17:58:37 GMT
- Title: Investigating the Behavior of Diffusion Models for Accelerating
Electronic Structure Calculations
- Authors: Daniel Rothchild, Andrew S. Rosen, Eric Taw, Connie Robinson, Joseph
E. Gonzalez, Aditi S. Krishnapriyan
- Abstract summary: Investigation driven by their potential to significantly accelerate electronic structure calculations using machine learning.
We show that the model learns about the first-order structure of the potential energy surface, and then later learns about higher-order structure.
For structure relaxations, the model finds geometries with 10x lower energy than those produced by a classical force field for small organic molecules.
- Score: 24.116064925926914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present an investigation into diffusion models for molecular generation,
with the aim of better understanding how their predictions compare to the
results of physics-based calculations. The investigation into these models is
driven by their potential to significantly accelerate electronic structure
calculations using machine learning, without requiring expensive
first-principles datasets for training interatomic potentials. We find that the
inference process of a popular diffusion model for de novo molecular generation
is divided into an exploration phase, where the model chooses the atomic
species, and a relaxation phase, where it adjusts the atomic coordinates to
find a low-energy geometry. As training proceeds, we show that the model
initially learns about the first-order structure of the potential energy
surface, and then later learns about higher-order structure. We also find that
the relaxation phase of the diffusion model can be re-purposed to sample the
Boltzmann distribution over conformations and to carry out structure
relaxations. For structure relaxations, the model finds geometries with ~10x
lower energy than those produced by a classical force field for small organic
molecules. Initializing a density functional theory (DFT) relaxation at the
diffusion-produced structures yields a >2x speedup to the DFT relaxation when
compared to initializing at structures relaxed with a classical force field.
Related papers
- Neural Message Passing Induced by Energy-Constrained Diffusion [79.9193447649011]
We propose an energy-constrained diffusion model as a principled interpretable framework for understanding the mechanism of MPNNs.
We show that the new model can yield promising performance for cases where the data structures are observed (as a graph), partially observed or completely unobserved.
arXiv Detail & Related papers (2024-09-13T17:54:41Z) - Molecule Design by Latent Prompt Transformer [76.2112075557233]
This work explores the challenging problem of molecule design by framing it as a conditional generative modeling task.
We propose a novel generative model comprising three components: (1) a latent vector with a learnable prior distribution; (2) a molecule generation model based on a causal Transformer, which uses the latent vector as a prompt; and (3) a property prediction model that predicts a molecule's target properties and/or constraint values using the latent prompt.
arXiv Detail & Related papers (2024-02-27T03:33:23Z) - Swallowing the Bitter Pill: Simplified Scalable Conformer Generation [12.341835649897886]
We present a novel way to predict molecular conformers through a simple formulation that sidesteps many of the equis of prior works and achieves state of the art results by using the advantages of scale.
We are able to radically simplify structure learning, and make it trivial to scale up the model sizes.
This model, called Molecular Conformer Fields (MCF), works by parameterizing conformer structures as functions that map elements from a molecular graph directly to their 3D location in space.
arXiv Detail & Related papers (2023-11-27T22:53:41Z) - Accelerating Inference in Molecular Diffusion Models with Latent Representations of Protein Structure [0.0]
Diffusion generative models operate directly on 3D molecular structures.
We present a novel GNN-based architecture for learning latent representations of molecular structure.
Our model achieves comparable performance to one with an all-atom protein representation while exhibiting a 3-fold reduction in inference time.
arXiv Detail & Related papers (2023-11-22T15:32:31Z) - Generative Design of inorganic compounds using deep diffusion language
models [0.6827423171182154]
We introduce a deep learning-based generative model for material composition and structure design.
Our pipeline first uses deep diffusion language models as the generator of compositions and then applies a template-based crystal structure prediction algorithm.
Based on the DFT calculation results, six new materials, including Ti2$HfO5, TaNbP, YMoN2, TaReO4, HfTiO2, and HfMnO2, with formation energy less than zero have been found.
arXiv Detail & Related papers (2023-09-30T19:46:19Z) - Molecule Design by Latent Space Energy-Based Modeling and Gradual
Distribution Shifting [53.44684898432997]
Generation of molecules with desired chemical and biological properties is critical for drug discovery.
We propose a probabilistic generative model to capture the joint distribution of molecules and their properties.
Our method achieves very strong performances on various molecule design tasks.
arXiv Detail & Related papers (2023-06-09T03:04:21Z) - Towards Predicting Equilibrium Distributions for Molecular Systems with
Deep Learning [60.02391969049972]
We introduce a novel deep learning framework, called Distributional Graphormer (DiG), in an attempt to predict the equilibrium distribution of molecular systems.
DiG employs deep neural networks to transform a simple distribution towards the equilibrium distribution, conditioned on a descriptor of a molecular system.
arXiv Detail & Related papers (2023-06-08T17:12:08Z) - Latent Diffusion Energy-Based Model for Interpretable Text Modeling [104.85356157724372]
We introduce a novel symbiosis between the diffusion models and latent space EBMs in a variational learning framework.
We develop a geometric clustering-based regularization jointly with the information bottleneck to further improve the quality of the learned latent space.
arXiv Detail & Related papers (2022-06-13T03:41:31Z) - A Score-based Geometric Model for Molecular Dynamics Simulations [33.158796937777886]
We propose a novel model called ScoreMD to estimate the gradient of the log density of molecular conformations.
With multiple architectural improvements, we outperforms state-of-the-art baselines on MD17 and isomers of C7O2H10.
This research provides new insights into the acceleration of new material and drug discovery.
arXiv Detail & Related papers (2022-04-19T05:13:46Z) - Learning Neural Generative Dynamics for Molecular Conformation
Generation [89.03173504444415]
We study how to generate molecule conformations (textiti.e., 3D structures) from a molecular graph.
We propose a novel probabilistic framework to generate valid and diverse conformations given a molecular graph.
arXiv Detail & Related papers (2021-02-20T03:17:58Z) - Embedded-physics machine learning for coarse-graining and collective
variable discovery without data [3.222802562733787]
We present a novel learning framework that consistently embeds underlying physics.
We propose a novel objective based on reverse Kullback-Leibler divergence that fully incorporates the available physics in the form of the atomistic force field.
We demonstrate the algorithmic advances in terms of predictive ability and the physical meaning of the revealed CVs for a bimodal potential energy function and the alanine dipeptide.
arXiv Detail & Related papers (2020-02-24T10:28:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.