DeepFoldit -- A Deep Reinforcement Learning Neural Network Folding
Proteins
- URL: http://arxiv.org/abs/2011.03442v1
- Date: Wed, 28 Oct 2020 16:05:42 GMT
- Title: DeepFoldit -- A Deep Reinforcement Learning Neural Network Folding
Proteins
- Authors: Dimitra N. Panou and Martin Reczko
- Abstract summary: We trained a deep reinforcement neural network called DeepFoldit to improve the score assigned to an unfolded protein.
Our approach combines the intuitive user interface of Foldit with the efficiency of deep reinforcement learning.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Despite considerable progress, ab initio protein structure prediction remains
suboptimal. A crowdsourcing approach is the online puzzle video game Foldit,
that provided several useful results that matched or even outperformed
algorithmically computed solutions. Using Foldit, the WeFold crowd had several
successful participations in the Critical Assessment of Techniques for Protein
Structure Prediction. Based on the recent Foldit standalone version, we trained
a deep reinforcement neural network called DeepFoldit to improve the score
assigned to an unfolded protein, using the Q-learning method with experience
replay. This paper is focused on model improvement through hyperparameter
tuning. We examined various implementations by examining different model
architectures and changing hyperparameter values to improve the accuracy of the
model. The new model hyper-parameters also improved its ability to generalize.
Initial results, from the latest implementation, show that given a set of small
unfolded training proteins, DeepFoldit learns action sequences that improve the
score both on the training set and on novel test proteins. Our approach
combines the intuitive user interface of Foldit with the efficiency of deep
reinforcement learning.
Related papers
- Deep Manifold Transformation for Protein Representation Learning [42.43017670985785]
We propose a new underlinedeep underlinemanifold underlinetrans approach for universal underlineprotein underlinerepresentation underlinelformation (DMTPRL)
It employs manifold learning strategies to improve the quality and adaptability of the learned embeddings.
Our proposed DMTPRL method outperforms state-of-the-art baselines on diverse downstream tasks across popular datasets.
arXiv Detail & Related papers (2024-01-12T18:38:14Z) - Target-aware Variational Auto-encoders for Ligand Generation with
Multimodal Protein Representation Learning [2.01243755755303]
We introduce TargetVAE, a target-aware auto-encoder that generates with high binding affinities to arbitrary protein targets.
This is the first effort to unify different representations of proteins into a single model that we name as Protein Multimodal Network (PMN)
arXiv Detail & Related papers (2023-08-02T12:08:17Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Boosting Convolutional Neural Networks' Protein Binding Site Prediction
Capacity Using SE(3)-invariant transformers, Transfer Learning and
Homology-based Augmentation [1.160208922584163]
Figuring out small binding sites in target proteins, in the resolution of either pocket or residue, is critical in real drugdiscovery scenarios.
Here we present a new computational method for binding site prediction that is relevant to real world applications.
arXiv Detail & Related papers (2023-02-20T05:02:40Z) - Learning Large-scale Neural Fields via Context Pruned Meta-Learning [60.93679437452872]
We introduce an efficient optimization-based meta-learning technique for large-scale neural field training.
We show how gradient re-scaling at meta-test time allows the learning of extremely high-quality neural fields.
Our framework is model-agnostic, intuitive, straightforward to implement, and shows significant reconstruction improvements for a wide range of signals.
arXiv Detail & Related papers (2023-02-01T17:32:16Z) - Reprogramming Pretrained Language Models for Protein Sequence
Representation Learning [68.75392232599654]
We propose Representation Learning via Dictionary Learning (R2DL), an end-to-end representation learning framework.
R2DL reprograms a pretrained English language model to learn the embeddings of protein sequences.
Our model can attain better accuracy and significantly improve the data efficiency by up to $105$ times over the baselines set by pretrained and standard supervised methods.
arXiv Detail & Related papers (2023-01-05T15:55:18Z) - Integration of Pre-trained Protein Language Models into Geometric Deep
Learning Networks [68.90692290665648]
We integrate knowledge learned by protein language models into several state-of-the-art geometric networks.
Our findings show an overall improvement of 20% over baselines.
Strong evidence indicates that the incorporation of protein language models' knowledge enhances geometric networks' capacity by a significant margin.
arXiv Detail & Related papers (2022-12-07T04:04:04Z) - Leveraging Angular Information Between Feature and Classifier for
Long-tailed Learning: A Prediction Reformulation Approach [90.77858044524544]
We reformulate the recognition probabilities through included angles without re-balancing the classifier weights.
Inspired by the performance improvement of the predictive form reformulation, we explore the different properties of this angular prediction.
Our method is able to obtain the best performance among peer methods without pretraining on CIFAR10/100-LT and ImageNet-LT.
arXiv Detail & Related papers (2022-12-03T07:52:48Z) - Analytically Tractable Bayesian Deep Q-Learning [0.0]
We adapt the temporal difference Q-learning framework to make it compatible with the tractable approximate Gaussian inference (TAGI)
We demonstrate that TAGI can reach a performance comparable to backpropagation-trained networks.
arXiv Detail & Related papers (2021-06-21T13:11:52Z) - EBM-Fold: Fully-Differentiable Protein Folding Powered by Energy-based
Models [53.17320541056843]
We propose a fully-differentiable approach for protein structure optimization, guided by a data-driven generative network.
Our EBM-Fold approach can efficiently produce high-quality decoys, compared against traditional Rosetta-based structure optimization routines.
arXiv Detail & Related papers (2021-05-11T03:40:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.