A unified cross-attention model for predicting antigen binding specificity to both HLA and TCR molecules
- URL: http://arxiv.org/abs/2405.06653v1
- Date: Mon, 8 Apr 2024 08:25:25 GMT
- Title: A unified cross-attention model for predicting antigen binding specificity to both HLA and TCR molecules
- Authors: Chenpeng Yu, Xing Fang, Hui Liu,
- Abstract summary: The binding affinity between antigens and HLA-I/TCR molecules plays a critical role in antigen presentation and T-cell activation.
Some computational methods have been developed to predict antigen-HLA or antigen-TCR binding specificity, but they focus solely on one task at a time.
We propose UnifyImmun, a unified cross-attention transformer model designed to simultaneously predict the binding of antigens to both HLA and TCR molecules.
- Score: 4.501817929699959
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: The immune checkpoint inhibitors have demonstrated promising clinical efficacy across various tumor types, yet the percentage of patients who benefit from them remains low. The binding affinity between antigens and HLA-I/TCR molecules plays a critical role in antigen presentation and T-cell activation. Some computational methods have been developed to predict antigen-HLA or antigen-TCR binding specificity, but they focus solely on one task at a time. In this paper, we propose UnifyImmun, a unified cross-attention transformer model designed to simultaneously predicts the binding of antigens to both HLA and TCR molecules, thereby providing more comprehensive evaluation of antigen immunogenicity. We devise a two-phase progressive training strategy that enables these two tasks to mutually reinforce each other, by compelling the encoders to extract more expressive features. To further enhance the model generalizability, we incorporate virtual adversarial training. Compared to over ten existing methods for predicting antigen-HLA and antigen-TCR binding, our method demonstrates better performance in both tasks. Notably, on a large-scale COVID-19 antigen-TCR binding test set, our method improves performance by at least 9% compared to the current state-of-the-art methods. The validation experiments on three clinical cohorts confirm that our approach effectively predicts immunotherapy response and clinical outcomes. Furthermore, the cross-attention scores reveal the amino acids sites critical for antigen binding to receptors. In essence, our approach marks a significant step towards comprehensive evaluation of antigen immunogenicity.
Related papers
- A large language model for predicting T cell receptor-antigen binding specificity [4.120928123714289]
We propose a Masked Language Model (MLM) to overcome limitations in model generalization.
Specifically, we randomly masked sequence segments and train tcrLM to infer the masked segment, thereby extract expressive feature from TCR sequences.
Our extensive experimental results demonstrate that tcrLM achieved AUC values of 0.937 and 0.933 on independent test sets and external validation sets.
arXiv Detail & Related papers (2024-06-24T08:36:40Z) - Antigen-Specific Antibody Design via Direct Energy-based Preference Optimization [51.28231365213679]
We tackle antigen-specific antibody sequence-structure co-design as an optimization problem towards specific preferences.
We propose direct energy-based preference optimization to guide the generation of antibodies with both rational structures and considerable binding affinities to given antigens.
arXiv Detail & Related papers (2024-03-25T09:41:49Z) - Evaluating Zero-Shot Scoring for In Vitro Antibody Binding Prediction
with Experimental Validation [0.08968838300743379]
We compare 8 common scoring paradigms based on open-source models to classify antibody designs as binders or non-binders.
Results show that existing methods struggle to detect binders, and performance is highly variable across antigens.
arXiv Detail & Related papers (2023-12-07T23:34:03Z) - Dense Feature Memory Augmented Transformers for COVID-19 Vaccination
Search Classification [60.49594822215981]
This paper presents a classification model for detecting COVID-19 vaccination related search queries.
We propose a novel approach of considering dense features as memory tokens that the model can attend to.
We show that this new modeling approach enables a significant improvement to the Vaccine Search Insights (VSI) task.
arXiv Detail & Related papers (2022-12-16T13:57:41Z) - xTrimoABFold: De novo Antibody Structure Prediction without MSA [77.47606749555686]
We develop a novel model named xTrimoABFold to predict antibody structure from antibody sequence.
The model was trained end-to-end on the antibody structures in PDB by minimizing the ensemble loss of domain-specific focal loss on CDR and the frame-aligned point loss.
arXiv Detail & Related papers (2022-11-30T09:26:08Z) - Incorporating Pre-training Paradigm for Antibody Sequence-Structure
Co-design [134.65287929316673]
Deep learning-based computational antibody design has attracted popular attention since it automatically mines the antibody patterns from data that could be complementary to human experiences.
The computational methods heavily rely on high-quality antibody structure data, which is quite limited.
Fortunately, there exists a large amount of sequence data of antibodies that can help model the CDR and alleviate the reliance on structure data.
arXiv Detail & Related papers (2022-10-26T15:31:36Z) - Deciphering antibody affinity maturation with language models and weakly
supervised learning [10.506336354512145]
We introduce AntiBERTy, a language model trained on 558M natural antibody sequences.
We find that within repertoires, our model clusters antibodies into trajectories resembling affinity maturation.
We show that models trained to predict highly redundant sequences under a multiple instance learning framework identify key binding residues in the process.
arXiv Detail & Related papers (2021-12-14T23:05:01Z) - Improved Drug-target Interaction Prediction with Intermolecular Graph
Transformer [98.8319016075089]
We propose a novel approach to model intermolecular information with a three-way Transformer-based architecture.
Intermolecular Graph Transformer (IGT) outperforms state-of-the-art approaches by 9.1% and 20.5% over the second best for binding activity and binding pose prediction respectively.
IGT exhibits promising drug screening ability against SARS-CoV-2 by identifying 83.1% active drugs that have been validated by wet-lab experiments with near-native predicted binding poses.
arXiv Detail & Related papers (2021-10-14T13:28:02Z) - Neural message passing for joint paratope-epitope prediction [0.0]
Antibodies are proteins in the immune system which bind to antigens to detect and neutralise them.
Prediction of binding sites in an antibody-antigen interaction are known as the paratope and, respectively, and are key to vaccine and synthetic antibody development.
arXiv Detail & Related papers (2021-05-31T16:37:55Z) - PaccMann$^{RL}$ on SARS-CoV-2: Designing antiviral candidates with
conditional generative models [2.0750380105212116]
With the fast development of COVID-19 into a global pandemic, scientists around the globe are desperately searching for effective antiviral therapeutic agents.
We propose a deep learning framework for conditional de novo design of antiviral candidate drugs tailored against given protein targets.
arXiv Detail & Related papers (2020-05-27T11:30:15Z) - Accelerating Antimicrobial Discovery with Controllable Deep Generative
Models and Molecular Dynamics [109.70543391923344]
CLaSS (Controlled Latent attribute Space Sampling) is an efficient computational method for attribute-controlled generation of molecules.
We screen the generated molecules for additional key attributes by using deep learning classifiers in conjunction with novel features derived from atomistic simulations.
The proposed approach is demonstrated for designing non-toxic antimicrobial peptides (AMPs) with strong broad-spectrum potency.
arXiv Detail & Related papers (2020-05-22T15:57:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.