Saturn: Sample-efficient Generative Molecular Design using Memory Manipulation
- URL: http://arxiv.org/abs/2405.17066v1
- Date: Mon, 27 May 2024 11:37:36 GMT
- Title: Saturn: Sample-efficient Generative Molecular Design using Memory Manipulation
- Authors: Jeff Guo, Philippe Schwaller,
- Abstract summary: Generative molecular design for drug discovery has recently achieved a wave of experimental validation.
The most important factor for downstream success is whether an in silico oracle is well correlated with the desired end-point.
We introduce Saturn, which leverages the Augmented Memory algorithm and demonstrates the first application of the Mamba architecture for generative molecular design.
- Score: 0.4037357056611557
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative molecular design for drug discovery has very recently achieved a wave of experimental validation, with language-based backbones being the most common architectures employed. The most important factor for downstream success is whether an in silico oracle is well correlated with the desired end-point. To this end, current methods use cheaper proxy oracles with higher throughput before evaluating the most promising subset with high-fidelity oracles. The ability to directly optimize high-fidelity oracles would greatly enhance generative design and be expected to improve hit rates. However, current models are not efficient enough to consider such a prospect, exemplifying the sample efficiency problem. In this work, we introduce Saturn, which leverages the Augmented Memory algorithm and demonstrates the first application of the Mamba architecture for generative molecular design. We elucidate how experience replay with data augmentation improves sample efficiency and how Mamba synergistically exploits this mechanism. Saturn outperforms 22 models on multi-parameter optimization tasks relevant to drug discovery and may possess sufficient sample efficiency to consider the prospect of directly optimizing high-fidelity oracles.
Related papers
- DocMamba: Efficient Document Pre-training with State Space Model [56.84200017560988]
We present DocMamba, a novel framework based on the state space model.
It is designed to reduce computational complexity to linear while preserving global modeling capabilities.
Experiments on the HRDoc confirm DocMamba's potential for length extrapolation.
arXiv Detail & Related papers (2024-09-18T11:34:28Z) - Retrosynthesis prediction enhanced by in-silico reaction data
augmentation [66.5643280109899]
We present RetroWISE, a framework that employs a base model inferred from real paired data to perform in-silico reaction generation and augmentation.
On three benchmark datasets, RetroWISE achieves the best overall performance against state-of-the-art models.
arXiv Detail & Related papers (2024-01-31T07:40:37Z) - Beam Enumeration: Probabilistic Explainability For Sample Efficient
Self-conditioned Molecular Design [0.4769602527256662]
Generative molecular design has moved from proof-of-concept to real-world applicability.
Key challenges in explainability and sample efficiency present opportunities to enhance generative design.
Beamion is generally applicable to any language-based molecular generative model.
arXiv Detail & Related papers (2023-09-25T08:43:13Z) - E^2VPT: An Effective and Efficient Approach for Visual Prompt Tuning [55.50908600818483]
Fine-tuning large-scale pretrained vision models for new tasks has become increasingly parameter-intensive.
We propose an Effective and Efficient Visual Prompt Tuning (E2VPT) approach for large-scale transformer-based model adaptation.
Our approach outperforms several state-of-the-art baselines on two benchmarks.
arXiv Detail & Related papers (2023-07-25T19:03:21Z) - Predicting Side Effect of Drug Molecules using Recurrent Neural Networks [2.089191490381739]
Failure to identify side effects before submission to regulatory groups can cost millions of dollars and months of additional research to the companies.
Prior approaches rely on complex model designs and excessive parameter counts for side effect predictions.
We propose a approach that allows for the utilization of simple neural networks, specifically the recurrent neural network, with a 98+% reduction in the number of required parameters.
arXiv Detail & Related papers (2023-05-17T16:56:19Z) - Augmented Memory: Capitalizing on Experience Replay to Accelerate De
Novo Molecular Design [0.0]
Molecular generative models should learn to satisfy a desired objective under minimal oracle evaluations.
We propose a novel algorithm called Augmented Memory that combines data augmentation with experience replay.
We show that scores obtained from oracle calls can be reused to update the model multiple times.
arXiv Detail & Related papers (2023-05-10T14:00:50Z) - IRGen: Generative Modeling for Image Retrieval [82.62022344988993]
In this paper, we present a novel methodology, reframing image retrieval as a variant of generative modeling.
We develop our model, dubbed IRGen, to address the technical challenge of converting an image into a concise sequence of semantic units.
Our model achieves state-of-the-art performance on three widely-used image retrieval benchmarks and two million-scale datasets.
arXiv Detail & Related papers (2023-03-17T17:07:36Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Learning Where To Look -- Generative NAS is Surprisingly Efficient [11.83842808044211]
We propose a generative model, paired with a surrogate predictor, that iteratively learns to generate samples from increasingly promising latent subspaces.
This approach leads to very effective and efficient architecture search, while keeping the query amount low.
arXiv Detail & Related papers (2022-03-16T16:27:11Z) - Forward Operator Estimation in Generative Models with Kernel Transfer
Operators [37.999297683250575]
We show that our formulation enables highly efficient distribution approximation and sampling, and offers surprisingly good empirical performance.
We also show that the algorithm also performs well in small sample size settings (in brain imaging)
arXiv Detail & Related papers (2021-12-01T06:54:31Z) - AdaLead: A simple and robust adaptive greedy search algorithm for
sequence design [55.41644538483948]
We develop an easy-to-directed, scalable, and robust evolutionary greedy algorithm (AdaLead)
AdaLead is a remarkably strong benchmark that out-competes more complex state of the art approaches in a variety of biologically motivated sequence design challenges.
arXiv Detail & Related papers (2020-10-05T16:40:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.