Lattice protein design using Bayesian learning
- URL: http://arxiv.org/abs/2003.06601v5
- Date: Mon, 14 Jun 2021 05:40:29 GMT
- Title: Lattice protein design using Bayesian learning
- Authors: Tomoei Takahashi, George Chikenji and Kei Tokita
- Abstract summary: Protein design is the inverse approach of the three-dimensional (3D) structure prediction for elucidating the relationship between the 3D structures and amino acid sequences.
Here, we propose a novel statistical mechanical design method using Bayesian learning, which can design lattice proteins without the exhaustive conformational search.
We find a strong linearity between the chemical potential of water and the number of surface residues, thereby revealing the relationship between protein structure and the effect of water molecules.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Protein design is the inverse approach of the three-dimensional (3D)
structure prediction for elucidating the relationship between the 3D structures
and amino acid sequences. In general, the computation of the protein design
involves a double loop: a loop for amino acid sequence changes and a loop for
an exhaustive conformational search for each amino acid sequence. Herein, we
propose a novel statistical mechanical design method using Bayesian learning,
which can design lattice proteins without the exhaustive conformational search.
We consider a thermodynamic hypothesis of the evolution of proteins and apply
it to the prior distribution of amino acid sequences. Furthermore, we take the
water effect into account in view of the grand canonical picture. As a result,
on applying the 2D lattice hydrophobic-polar (HP) model, our design method
successfully finds an amino acid sequence for which the target conformation has
a unique ground state. However, the performance was not as good for the 3D
lattice HP models compared to the 2D models. The performance of the 3D model
improves on using a 20-letter lattice proteins. Furthermore, we find a strong
linearity between the chemical potential of water and the number of surface
residues, thereby revealing the relationship between protein structure and the
effect of water molecules. The advantage of our method is that it greatly
reduces computation time, because it does not require long calculations for the
partition function corresponding to an exhaustive conformational search. As our
method uses a general form of Bayesian learning and statistical mechanics and
is not limited to lattice proteins, the results presented here elucidate some
heuristics used successfully in previous protein design methods.
Related papers
- CrysFormer: Protein Structure Prediction via 3d Patterson Maps and
Partial Structure Attention [7.716601082662128]
A protein's three-dimensional structure often poses nontrivial computation costs.
We propose the first transformer-based model that directly utilizes protein crystallography and partial structure information.
We demonstrate our method, dubbed textttCrysFormer, can achieve accurate predictions, based on a much smaller dataset size and with reduced computation costs.
arXiv Detail & Related papers (2023-10-05T21:10:22Z) - Joint Design of Protein Sequence and Structure based on Motifs [11.731131799546489]
We propose GeoPro, a method to design protein backbone structure and sequence jointly.
GeoPro is powered by an equivariant encoder for three-dimensional (3D) backbone structure and a protein sequence decoder guided by 3D geometry.
Our method discovers novel $beta$-lactamases and myoglobins which are not present in protein data bank (PDB) and UniProt.
arXiv Detail & Related papers (2023-10-04T03:07:03Z) - A Latent Diffusion Model for Protein Structure Generation [50.74232632854264]
We propose a latent diffusion model that can reduce the complexity of protein modeling.
We show that our method can effectively generate novel protein backbone structures with high designability and efficiency.
arXiv Detail & Related papers (2023-05-06T19:10:19Z) - EquiPocket: an E(3)-Equivariant Geometric Graph Neural Network for Ligand Binding Site Prediction [49.674494450107005]
Predicting the binding sites of target proteins plays a fundamental role in drug discovery.
Most existing deep-learning methods consider a protein as a 3D image by spatially clustering its atoms into voxels.
This work proposes EquiPocket, an E(3)-equivariant Graph Neural Network (GNN) for binding site prediction.
arXiv Detail & Related papers (2023-02-23T17:18:26Z) - Learning Geometrically Disentangled Representations of Protein Folding
Simulations [72.03095377508856]
This work focuses on learning a generative neural network on a structural ensemble of a drug-target protein.
Model tasks involve characterizing the distinct structural fluctuations of the protein bound to various drug molecules.
Results show that our geometric learning-based method enjoys both accuracy and efficiency for generating complex structural variations.
arXiv Detail & Related papers (2022-05-20T19:38:00Z) - MAS2HP: A Multi Agent System to predict protein structure in 2D HP model [0.0]
We propose a new approach for protein structure prediction by using agent-based modeling (ABM) in two dimensional hydrophobic-hydrophilic model.
We have tested this algorithm on several benchmark sequences ranging from 20 to 50-mers in two dimensional Hydrophobic-Hydrophilic lattice models.
arXiv Detail & Related papers (2022-05-11T05:17:47Z) - Independent SE(3)-Equivariant Models for End-to-End Rigid Protein
Docking [57.2037357017652]
We tackle rigid body protein-protein docking, i.e., computationally predicting the 3D structure of a protein-protein complex from the individual unbound structures.
We design a novel pairwise-independent SE(3)-equivariant graph matching network to predict the rotation and translation to place one of the proteins at the right docked position.
Our model, named EquiDock, approximates the binding pockets and predicts the docking poses using keypoint matching and alignment.
arXiv Detail & Related papers (2021-11-15T18:46:37Z) - G-VAE, a Geometric Convolutional VAE for ProteinStructure Generation [41.66010308405784]
We introduce a joint geometric-neural networks approach for comparing, deforming and generating 3D protein structures.
Our method is able to generate plausible structures, different from the structures in the training data.
arXiv Detail & Related papers (2021-06-22T16:52:48Z) - Deep Learning of High-Order Interactions for Protein Interface
Prediction [58.164371994210406]
We propose to formulate the protein interface prediction as a 2D dense prediction problem.
We represent proteins as graphs and employ graph neural networks to learn node features.
We incorporate high-order pairwise interactions to generate a 3D tensor containing different pairwise interactions.
arXiv Detail & Related papers (2020-07-18T05:39:35Z) - BERTology Meets Biology: Interpreting Attention in Protein Language
Models [124.8966298974842]
We demonstrate methods for analyzing protein Transformer models through the lens of attention.
We show that attention captures the folding structure of proteins, connecting amino acids that are far apart in the underlying sequence, but spatially close in the three-dimensional structure.
We also present a three-dimensional visualization of the interaction between attention and protein structure.
arXiv Detail & Related papers (2020-06-26T21:50:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.