State-of-the-Art Augmented NLP Transformer models for direct and
single-step retrosynthesis
- URL: http://arxiv.org/abs/2003.02804v2
- Date: Tue, 22 Sep 2020 14:23:59 GMT
- Title: State-of-the-Art Augmented NLP Transformer models for direct and
single-step retrosynthesis
- Authors: Igor V. Tetko, Pavel Karpov, Ruud Van Deursen, Guillaume Godin
- Abstract summary: We investigated the effect of different training scenarios on predicting retrosynthesis of chemical compounds.
Data augmentation, which is a powerful method used in image processing, eliminated the effect of data memorization by neural networks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We investigated the effect of different training scenarios on predicting the
(retro)synthesis of chemical compounds using a text-like representation of
chemical reactions (SMILES) and Natural Language Processing neural network
Transformer architecture. We showed that data augmentation, which is a powerful
method used in image processing, eliminated the effect of data memorization by
neural networks, and improved their performance for the prediction of new
sequences. This effect was observed when augmentation was used simultaneously
for input and the target data simultaneously. The top-5 accuracy was 84.8% for
the prediction of the largest fragment (thus identifying principal
transformation for classical retro-synthesis) for the USPTO-50k test dataset
and was achieved by a combination of SMILES augmentation and a beam search
algorithm. The same approach provided significantly better results for the
prediction of direct reactions from the single-step USPTO-MIT test set. Our
model achieved 90.6% top-1 and 96.1% top-5 accuracy for its challenging mixed
set and 97% top-5 accuracy for the USPTO-MIT separated set. It also
significantly improved results for USPTO-full set single-step retrosynthesis
for both top-1 and top-10 accuracies. The appearance frequency of the most
abundantly generated SMILES was well correlated with the prediction outcome and
can be used as a measure of the quality of reaction prediction.
Related papers
- PredFormer: Transformers Are Effective Spatial-Temporal Predictive Learners [65.93130697098658]
This paper proposes PredFormer, a pure transformer-based framework for predictive learning.
With its recurrent-free, transformer-based design, PredFormer is both simple and efficient.
experiments on synthetic and real-world datasets demonstrate that PredFormer achieves state-the-art performance.
arXiv Detail & Related papers (2024-10-07T03:52:06Z) - Retrosynthesis prediction enhanced by in-silico reaction data
augmentation [66.5643280109899]
We present RetroWISE, a framework that employs a base model inferred from real paired data to perform in-silico reaction generation and augmentation.
On three benchmark datasets, RetroWISE achieves the best overall performance against state-of-the-art models.
arXiv Detail & Related papers (2024-01-31T07:40:37Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Comparative Analysis of Epileptic Seizure Prediction: Exploring Diverse
Pre-Processing Techniques and Machine Learning Models [0.0]
We present a comparative analysis of five machine learning models for the prediction of epileptic seizures using EEG data.
The results of our analysis demonstrate the performance of each model in terms of accuracy.
The ET model exhibited the best performance with an accuracy of 99.29%.
arXiv Detail & Related papers (2023-08-06T08:50:08Z) - Retrosynthesis Prediction with Local Template Retrieval [112.23386062396622]
Retrosynthesis, which predicts the reactants of a given target molecule, is an essential task for drug discovery.
In this work, we introduce RetroKNN, a local reaction template retrieval method.
We conduct comprehensive experiments on two widely used benchmarks, the USPTO-50K and USPTO-MIT.
arXiv Detail & Related papers (2023-06-07T03:38:03Z) - G-MATT: Single-step Retrosynthesis Prediction using Molecular Grammar
Tree Transformer [0.0]
We propose a chemistry-aware retrosynthesis prediction framework that combines powerful data-driven models with prior domain knowledge.
The proposed framework, grammar-based molecular attention tree transformer (G-MATT), achieves significant performance improvements compared to baseline retrosynthesis models.
arXiv Detail & Related papers (2023-05-04T21:04:19Z) - Unassisted Noise Reduction of Chemical Reaction Data Sets [59.127921057012564]
We propose a machine learning-based, unassisted approach to remove chemically wrong entries from data sets.
Our results show an improved prediction quality for models trained on the cleaned and balanced data sets.
arXiv Detail & Related papers (2021-02-02T09:34:34Z) - A Generative Model to Synthesize EEG Data for Epileptic Seizure
Prediction [3.8271082752302137]
This paper proposes a deep convolutional generative adversarial network to generate synthetic EEG samples.
We use two methods to validate synthesized data namely, one-class SVM and a new proposal which we refer to as convolutional epileptic seizure predictor (CESP)
Our results show that CESP model achieves sensitivity of 78.11% and 88.21%, and FPR of 0.27/h and 0.14/h for training on synthesized data.
arXiv Detail & Related papers (2020-12-01T12:00:36Z) - Towards a Competitive End-to-End Speech Recognition for CHiME-6 Dinner
Party Transcription [73.66530509749305]
In this paper, we argue that, even in difficult cases, some end-to-end approaches show performance close to the hybrid baseline.
We experimentally compare and analyze CTC-Attention versus RNN-Transducer approaches along with RNN versus Transformer architectures.
Our best end-to-end model based on RNN-Transducer, together with improved beam search, reaches quality by only 3.8% WER abs. worse than the LF-MMI TDNN-F CHiME-6 Challenge baseline.
arXiv Detail & Related papers (2020-04-22T19:08:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.