Generating stable molecules using imitation and reinforcement learning
- URL: http://arxiv.org/abs/2107.05007v1
- Date: Sun, 11 Jul 2021 10:18:19 GMT
- Title: Generating stable molecules using imitation and reinforcement learning
- Authors: S{\o}ren Ager Meldgaard, Jonas K\"ohler, Henrik Lund Mortensen,
Mads-Peter V. Christiansen, Frank No\'e, Bj{\o}rk Hammer
- Abstract summary: We propose a reinforcement learning approach for generating molecules in cartesian coordinates.
We learn basic chemical rules from imitation learning on the GDB-11 database to create an initial model.
We then deploy multiple copies of the model conditioned on a specific stoichiometry in a reinforcement learning setting.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Chemical space is routinely explored by machine learning methods to discover
interesting molecules, before time-consuming experimental synthesizing is
attempted. However, these methods often rely on a graph representation,
ignoring 3D information necessary for determining the stability of the
molecules. We propose a reinforcement learning approach for generating
molecules in cartesian coordinates allowing for quantum chemical prediction of
the stability. To improve sample-efficiency we learn basic chemical rules from
imitation learning on the GDB-11 database to create an initial model applicable
for all stoichiometries. We then deploy multiple copies of the model
conditioned on a specific stoichiometry in a reinforcement learning setting.
The models correctly identify low energy molecules in the database and produce
novel isomers not found in the training set. Finally, we apply the model to
larger molecules to show how reinforcement learning further refines the
imitation learning model in domains far from the training data.
Related papers
- Active Learning Enables Extrapolation in Molecular Generative Models [11.234291560423943]
We create an active-learning, closed-loop molecule generation pipeline for molecular generative models.
Compared against other generative model approaches, only our active learning approach generates molecules with properties that extrapolate beyond the training data.
The proportion of stable molecules generated is 3.5x higher than the next-best model.
arXiv Detail & Related papers (2025-01-03T19:07:06Z) - Pre-trained Molecular Language Models with Random Functional Group Masking [54.900360309677794]
We propose a SMILES-based underlineem Molecular underlineem Language underlineem Model, which randomly masking SMILES subsequences corresponding to specific molecular atoms.
This technique aims to compel the model to better infer molecular structures and properties, thus enhancing its predictive capabilities.
arXiv Detail & Related papers (2024-11-03T01:56:15Z) - Conditional Synthesis of 3D Molecules with Time Correction Sampler [58.0834973489875]
Time-Aware Conditional Synthesis (TACS) is a novel approach to conditional generation on diffusion models.
It integrates adaptively controlled plug-and-play "online" guidance into a diffusion model, driving samples toward the desired properties.
arXiv Detail & Related papers (2024-11-01T12:59:25Z) - Molecule Design by Latent Space Energy-Based Modeling and Gradual
Distribution Shifting [53.44684898432997]
Generation of molecules with desired chemical and biological properties is critical for drug discovery.
We propose a probabilistic generative model to capture the joint distribution of molecules and their properties.
Our method achieves very strong performances on various molecule design tasks.
arXiv Detail & Related papers (2023-06-09T03:04:21Z) - Improving Molecular Pretraining with Complementary Featurizations [20.86159731100242]
Molecular pretraining is a paradigm to solve a variety of tasks in computational chemistry and drug discovery.
We show that different featurization techniques convey chemical information differently.
We propose a simple and effective MOlecular pretraining framework with COmplementary featurizations (MOCO)
arXiv Detail & Related papers (2022-09-29T21:11:09Z) - A Molecular Multimodal Foundation Model Associating Molecule Graphs with
Natural Language [63.60376252491507]
We propose a molecular multimodal foundation model which is pretrained from molecular graphs and their semantically related textual data.
We believe that our model would have a broad impact on AI-empowered fields across disciplines such as biology, chemistry, materials, environment, and medicine.
arXiv Detail & Related papers (2022-09-12T00:56:57Z) - Learning Neural Generative Dynamics for Molecular Conformation
Generation [89.03173504444415]
We study how to generate molecule conformations (textiti.e., 3D structures) from a molecular graph.
We propose a novel probabilistic framework to generate valid and diverse conformations given a molecular graph.
arXiv Detail & Related papers (2021-02-20T03:17:58Z) - Learning Latent Space Energy-Based Prior Model for Molecule Generation [59.875533935578375]
We learn latent space energy-based prior model with SMILES representation for molecule modeling.
Our method is able to generate molecules with validity and uniqueness competitive with state-of-the-art models.
arXiv Detail & Related papers (2020-10-19T09:34:20Z) - Learning a Continuous Representation of 3D Molecular Structures with
Deep Generative Models [0.0]
Generative models are an entirely different approach that learn to represent and optimize molecules in a continuous latent space.
We describe deep generative models of three dimensional molecular structures using atomic density grids.
We are also able to sample diverse sets of molecules based on a given input compound to increase the probability of creating valid, drug-like molecules.
arXiv Detail & Related papers (2020-10-17T01:15:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.