TCR-EML: Explainable Model Layers for TCR-pMHC Prediction
- URL: http://arxiv.org/abs/2510.04377v1
- Date: Sun, 05 Oct 2025 21:47:48 GMT
- Title: TCR-EML: Explainable Model Layers for TCR-pMHC Prediction
- Authors: Jiarui Li, Zixiang Yin, Zhengming Ding, Samuel J. Landry, Ramgopal R. Mettu,
- Abstract summary: T cell receptor (TCR) recognition of peptide-MHC (pMHC) complexes is a central component of adaptive immunity.<n>We propose explainable model layers (TCR-EML) that can be incorporated into protein-language model backbones for TCR-pMHC modeling.
- Score: 31.62439146845817
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: T cell receptor (TCR) recognition of peptide-MHC (pMHC) complexes is a central component of adaptive immunity, with implications for vaccine design, cancer immunotherapy, and autoimmune disease. While recent advances in machine learning have improved prediction of TCR-pMHC binding, the most effective approaches are black-box transformer models that cannot provide a rationale for predictions. Post-hoc explanation methods can provide insight with respect to the input but do not explicitly model biochemical mechanisms (e.g. known binding regions), as in TCR-pMHC binding. ``Explain-by-design'' models (i.e., with architectural components that can be examined directly after training) have been explored in other domains, but have not been used for TCR-pMHC binding. We propose explainable model layers (TCR-EML) that can be incorporated into protein-language model backbones for TCR-pMHC modeling. Our approach uses prototype layers for amino acid residue contacts drawn from known TCR-pMHC binding mechanisms, enabling high-quality explanations for predicted TCR-pMHC binding. Experiments of our proposed method on large-scale datasets demonstrate competitive predictive accuracy and generalization, and evaluation on the TCR-XAI benchmark demonstrates improved explainability compared with existing approaches.
Related papers
- LGE-Guided Cross-Modality Contrastive Learning for Gadolinium-Free Cardiomyopathy Screening in Cine CMR [51.11296719862485]
We propose a Contrastive Learning and Cross-Modal alignment framework for gadolinium-free cardiomyopathy screening using cine CMR sequences.<n>By aligning the latent spaces of cine CMR and Late Gadolinium Enhancement (LGE) sequences, our model encodes fibrosis-specific pathology into cine CMR embeddings.
arXiv Detail & Related papers (2025-08-23T07:21:23Z) - Generation of structure-guided pMHC-I libraries using Diffusion Models [0.0]
We introduce a structure-guided benchmark of pMHC-I peptides designed using diffusion models conditioned on crystal distances.<n>This benchmark is independent of previously characterized peptides yet reproduces canonical anchor residue preferences.<n>We demonstrate that state-of-the-art sequence-based predictors perform poorly at recognizing the binding potential of these structurally stable designs.
arXiv Detail & Related papers (2025-07-11T08:29:18Z) - Quantifying Cross-Attention Interaction in Transformers for Interpreting TCR-pMHC Binding [33.31929328499189]
Quantifying Cross-Attention Interaction (QCAI) is a new post-hoc method designed to interpret the cross-attention mechanisms in transformer decoders.<n>QCAI achieves state-of-the-art performance on both interpretability and prediction accuracy under the TCR-XAI benchmark.
arXiv Detail & Related papers (2025-07-03T22:18:54Z) - Enhancing TCR-Peptide Interaction Prediction with Pretrained Language Models and Molecular Representations [0.39945675027960637]
We present LANTERN, a deep learning framework that combines large-scale protein language models with chemical representations of peptides.<n>Our model demonstrates superior performance, particularly in zero-shot and few-shot learning scenarios.<n>These results highlight the potential of LANTERN to advance TCR-pMHC binding prediction and support the development of personalized immunotherapies.
arXiv Detail & Related papers (2025-04-22T20:22:34Z) - MCTS-RAG: Enhancing Retrieval-Augmented Generation with Monte Carlo Tree Search [61.11836311160951]
We introduce MCTS-RAG, a novel approach that enhances the reasoning capabilities of small language models on knowledge-intensive tasks.<n>Unlike standard RAG methods, which typically retrieve information independently from reasoning, MCTS-RAG combines structured reasoning with adaptive retrieval.<n>This integrated approach enhances decision-making, reduces hallucinations, and ensures improved factual accuracy and response consistency.
arXiv Detail & Related papers (2025-03-26T17:46:08Z) - T-cell receptor specificity landscape revealed through de novo peptide design [2.1892916695245224]
We introduce a computational approach to predict interactions between T-cell receptors (TCRs) and peptides presented on Major Histocompatibility Complexes (MHCs)<n>We use HERMES, a structure-based, physics-guided machine learning model trained on the protein universe to predict amino acid preferences based on local structural environments.<n>Our approach provides a platform for immunogenic peptide and neoantigen design, as well as for evaluating TCR specificity, offering a computational framework to inform design of engineered T-cell therapies and vaccines.
arXiv Detail & Related papers (2025-03-01T22:45:19Z) - Fast Context-Biasing for CTC and Transducer ASR models with CTC-based Word Spotter [57.64003871384959]
This work presents a new approach to fast context-biasing with CTC-based Word Spotter.
The proposed method matches CTC log-probabilities against a compact context graph to detect potential context-biasing candidates.
The results demonstrate a significant acceleration of the context-biasing recognition with a simultaneous improvement in F-score and WER.
arXiv Detail & Related papers (2024-06-11T09:37:52Z) - From Cloze to Comprehension: Retrofitting Pre-trained Masked Language
Model to Pre-trained Machine Reader [130.45769668885487]
Pre-trained Machine Reader (PMR) is a novel method for retrofitting masked language models (MLMs) to pre-trained machine reading comprehension (MRC) models without acquiring labeled data.
To build the proposed PMR, we constructed a large volume of general-purpose and high-quality MRC-style training data.
PMR has the potential to serve as a unified model for tackling various extraction and classification tasks in the MRC formulation.
arXiv Detail & Related papers (2022-12-09T10:21:56Z) - Attention-aware contrastive learning for predicting T cell
receptor-antigen binding specificity [7.365824008999903]
It has been verified that only a small fraction of the neoantigens presented by MHC class I molecules on the cell surface can elicit T cells.
We propose an attentive-mask contrastive learning model, ATMTCR, for inferring TCR-antigen binding specificity.
arXiv Detail & Related papers (2022-05-17T10:53:32Z) - Confidence-guided Lesion Mask-based Simultaneous Synthesis of Anatomic
and Molecular MR Images in Patients with Post-treatment Malignant Gliomas [65.64363834322333]
Confidence Guided SAMR (CG-SAMR) synthesizes data from lesion information to multi-modal anatomic sequences.
module guides the synthesis based on confidence measure about the intermediate results.
experiments on real clinical data demonstrate that the proposed model can perform better than the state-of-theart synthesis methods.
arXiv Detail & Related papers (2020-08-06T20:20:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.