FFF: Fragments-Guided Flexible Fitting for Building Complete Protein
Structures
- URL: http://arxiv.org/abs/2308.03654v1
- Date: Mon, 7 Aug 2023 15:10:21 GMT
- Title: FFF: Fragments-Guided Flexible Fitting for Building Complete Protein
Structures
- Authors: Weijie Chen, Xinyan Wang, Yuhang Wang
- Abstract summary: We propose a new method named FFF that bridges protein structure prediction and protein structure recognition with flexible fitting.
First, a multi-level recognition network is used to capture various structural features from the input 3D cryo-EM map.
Next, protein structural fragments are generated using pseudo peptide vectors and a protein sequence alignment method based on these extracted features.
- Score: 10.682516227941592
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Cryo-electron microscopy (cryo-EM) is a technique for reconstructing the
3-dimensional (3D) structure of biomolecules (especially large protein
complexes and molecular assemblies). As the resolution increases to the
near-atomic scale, building protein structures de novo from cryo-EM maps
becomes possible. Recently, recognition-based de novo building methods have
shown the potential to streamline this process. However, it cannot build a
complete structure due to the low signal-to-noise ratio (SNR) problem. At the
same time, AlphaFold has led to a great breakthrough in predicting protein
structures. This has inspired us to combine fragment recognition and structure
prediction methods to build a complete structure. In this paper, we propose a
new method named FFF that bridges protein structure prediction and protein
structure recognition with flexible fitting. First, a multi-level recognition
network is used to capture various structural features from the input 3D
cryo-EM map. Next, protein structural fragments are generated using pseudo
peptide vectors and a protein sequence alignment method based on these
extracted features. Finally, a complete structural model is constructed using
the predicted protein fragments via flexible fitting. Based on our benchmark
tests, FFF outperforms the baseline methods for building complete protein
structures.
Related papers
- CPE-Pro: A Structure-Sensitive Deep Learning Method for Protein Representation and Origin Evaluation [7.161099050722313]
We develop a structure-sensitive supervised deep learning model, Crystal vs Predicted Evaluator for Protein Structure (CPE-Pro)
CPE-Pro learns the structural information of proteins and captures inter-structural differences to achieve accurate traceability on four data classes.
We utilize Foldseek to encode protein structures into "structure-sequences" and trained a protein Structural Sequence Language Model, SSLM.
arXiv Detail & Related papers (2024-10-21T02:21:56Z) - Sequence-Augmented SE(3)-Flow Matching For Conditional Protein Backbone Generation [55.93511121486321]
We introduce FoldFlow-2, a novel sequence-conditioned flow matching model for protein structure generation.
We train FoldFlow-2 at scale on a new dataset that is an order of magnitude larger than PDB datasets of prior works.
We empirically observe that FoldFlow-2 outperforms previous state-of-the-art protein structure-based generative models.
arXiv Detail & Related papers (2024-05-30T17:53:50Z) - cryoSPHERE: Single-particle heterogeneous reconstruction from cryo EM [9.887133861477233]
Single-particle cryo-electron microscopy (cryo EM) is a powerful tool for capturing a large number of images of a given protein.
In this paper, we introduce cryoSPHERE, a deep learning method that takes as input a nominal protein structure.
arXiv Detail & Related papers (2024-05-29T15:12:19Z) - A Protein Structure Prediction Approach Leveraging Transformer and CNN
Integration [4.909112037834705]
This paper adopts a two-dimensional fusion deep neural network model, DstruCCN, which uses Convolutional Neural Networks (CCN) and a supervised Transformer protein language model for single-sequence protein structure prediction.
The training features of the two are combined to predict the protein Transformer binding site matrix, and then the three-dimensional structure is reconstructed using energy minimization.
arXiv Detail & Related papers (2024-02-29T12:24:20Z) - CrysFormer: Protein Structure Prediction via 3d Patterson Maps and
Partial Structure Attention [7.716601082662128]
A protein's three-dimensional structure often poses nontrivial computation costs.
We propose the first transformer-based model that directly utilizes protein crystallography and partial structure information.
We demonstrate our method, dubbed textttCrysFormer, can achieve accurate predictions, based on a much smaller dataset size and with reduced computation costs.
arXiv Detail & Related papers (2023-10-05T21:10:22Z) - Structure-informed Language Models Are Protein Designers [69.70134899296912]
We present LM-Design, a generic approach to reprogramming sequence-based protein language models (pLMs)
We conduct a structural surgery on pLMs, where a lightweight structural adapter is implanted into pLMs and endows it with structural awareness.
Experiments show that our approach outperforms the state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2023-02-03T10:49:52Z) - State-specific protein-ligand complex structure prediction with a
multi-scale deep generative model [68.28309982199902]
We present NeuralPLexer, a computational approach that can directly predict protein-ligand complex structures.
Our study suggests that a data-driven approach can capture the structural cooperativity between proteins and small molecules, showing promise in accelerating the design of enzymes, drug molecules, and beyond.
arXiv Detail & Related papers (2022-09-30T01:46:38Z) - Learning Geometrically Disentangled Representations of Protein Folding
Simulations [72.03095377508856]
This work focuses on learning a generative neural network on a structural ensemble of a drug-target protein.
Model tasks involve characterizing the distinct structural fluctuations of the protein bound to various drug molecules.
Results show that our geometric learning-based method enjoys both accuracy and efficiency for generating complex structural variations.
arXiv Detail & Related papers (2022-05-20T19:38:00Z) - Transfer Learning for Protein Structure Classification at Low Resolution [124.5573289131546]
We show that it is possible to make accurate ($geq$80%) predictions of protein class and architecture from structures determined at low ($leq$3A) resolution.
We provide proof of concept for high-speed, low-cost protein structure classification at low resolution, and a basis for extension to prediction of function.
arXiv Detail & Related papers (2020-08-11T15:01:32Z) - BERTology Meets Biology: Interpreting Attention in Protein Language
Models [124.8966298974842]
We demonstrate methods for analyzing protein Transformer models through the lens of attention.
We show that attention captures the folding structure of proteins, connecting amino acids that are far apart in the underlying sequence, but spatially close in the three-dimensional structure.
We also present a three-dimensional visualization of the interaction between attention and protein structure.
arXiv Detail & Related papers (2020-06-26T21:50:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.