Encoder-Free ECG-Language Models
- URL: http://arxiv.org/abs/2601.18798v1
- Date: Mon, 05 Jan 2026 08:38:39 GMT
- Title: Encoder-Free ECG-Language Models
- Authors: William Han, Tony Chen, Chaojing Duan, Xiaoyu Song, Yihang Yao, Yuzhe Yang, Michael A. Rosenberg, Emerson Liu, Ding Zhao,
- Abstract summary: ECG-Language Models (ELMs) extend recent progress in automated ECG interpretation.<n>We introduce ELF, an encoder-free ELM that replaces the ECG encoder with a single projection layer trained jointly with the LLM.<n>Across five datasets, ELF matches or exceeds state-of-the-art ELMs that use far more complex encoders and training pipelines.
- Score: 30.25983036594147
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: ECG-Language Models (ELMs) extend recent progress in Multimodal Large Language Models (MLLMs) to automated ECG interpretation. However, most ELMs follow Vision-Language Model (VLM) designs and depend on pretrained ECG encoders, adding architectural and training complexity. Inspired by encoder-free VLMs, we introduce ELF, an encoder-free ELM that replaces the ECG encoder with a single projection layer trained jointly with the LLM. Across five datasets, ELF matches or exceeds state-of-the-art ELMs that use far more complex encoders and training pipelines. We also test whether adding architectural biases to ELF improves performance and find that the single linear projection remains competitive. Finally, we show that ELF, and potentially other ELMs, often rely more on benchmark artifacts and language priors than ECG-derived information, highlighting limitations in current evaluation practices and ELM design. All data and code is available at https://github.com/willxxy/ECG-Bench.
Related papers
- A Hybrid Approach for EMF Code Generation:Code Templates Meet Large Language Models [7.176489635263981]
iEcoreGen is a hybrid approach that integrates Eclipse Modeling Framework (EMF) and LLMs.<n>iEcoreGen decomposes requirements to derive operation specifications, uses EMF's template-based generator to produce initial Java code, and serializes specifications into docstrings.
arXiv Detail & Related papers (2025-12-05T07:46:51Z) - Retrieval-Augmented Generation for Electrocardiogram-Language Models [27.75347676208195]
generative Electrocardiogram-Language Models (ELMs) can produce textual responses conditioned on ECG signals.<n>Retrieval-Augmented Generation (RAG) helps reduce hallucinations and improve natural language generation (NLG)
arXiv Detail & Related papers (2025-09-30T20:32:34Z) - ECG-aBcDe: Overcoming Model Dependence, Encoding ECG into a Universal Language for Any LLM [7.632459372363093]
Large Language Models (LLMs) hold significant promise for electrocardiogram (ECG) analysis.<n>Current methods suffer from model-specific ECG encoders, hindering transfer across LLMs.<n>We introduce ECG-aBcDe, a novel encoding method that transforms ECG signals into a universal ECG language readily interpretable by any LLM.
arXiv Detail & Related papers (2025-09-16T03:41:02Z) - ECG-Byte: A Tokenizer for End-to-End Generative Electrocardiogram Language Modeling [20.484166589932702]
Large Language Models (LLMs) have demonstrated exceptional versatility across domains, including applications to electrocardiograms (ECGs)<n>We propose ECG-Byte, an adapted byte pair encoding (BPE) tokenizer pipeline for autoregressive language modeling of ECGs.<n>We achieve competitive NLG performance while training 3 times faster and using just 48% of the data required by traditional two-stage methods.
arXiv Detail & Related papers (2024-12-18T22:13:21Z) - Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training [48.455597568212944]
We present Mono-InternVL, a novel monolithic MLLM that seamlessly integrates a set of visual experts via a multimodal mixture-of-experts structure.<n>In particular, EViP is designed as a progressive learning process for visual experts, which aims to fully exploit the visual knowledge from noisy data to high-quality data.
arXiv Detail & Related papers (2024-10-10T17:59:22Z) - LLMAEL: Large Language Models are Good Context Augmenters for Entity Linking [46.577071330549984]
Specialized entity linking (EL) models are well-trained at mapping mentions to unique knowledge base (KB) entities.<n>Extensively pre-trained large language models (LLMs) possess broader knowledge of uncommon entities.<n>We introduce LLM-Augmented Entity Linking (LLMAEL), the first framework to enhance specialized EL models with LLM data augmentation.
arXiv Detail & Related papers (2024-07-04T15:55:13Z) - DARG: Dynamic Evaluation of Large Language Models via Adaptive Reasoning Graph [70.79413606968814]
We introduce Dynamic Evaluation of LLMs via Adaptive Reasoning Graph Evolvement (DARG) to dynamically extend current benchmarks with controlled complexity and diversity.
Specifically, we first extract the reasoning graphs of data points in current benchmarks and then perturb the reasoning graphs to generate novel testing data.
Such newly generated test samples can have different levels of complexity while maintaining linguistic diversity similar to the original benchmarks.
arXiv Detail & Related papers (2024-06-25T04:27:53Z) - InfiBench: Evaluating the Question-Answering Capabilities of Code Large Language Models [56.723509505549536]
InfiBench is the first large-scale freeform question-answering (QA) benchmark for code to our knowledge.
It comprises 234 carefully selected high-quality Stack Overflow questions that span across 15 programming languages.
We conduct a systematic evaluation for over 100 latest code LLMs on InfiBench, leading to a series of novel and insightful findings.
arXiv Detail & Related papers (2024-03-11T02:06:30Z) - LLM4EDA: Emerging Progress in Large Language Models for Electronic
Design Automation [74.7163199054881]
Large Language Models (LLMs) have demonstrated their capability in context understanding, logic reasoning and answer generation.
We present a systematic study on the application of LLMs in the EDA field.
We highlight the future research direction, focusing on applying LLMs in logic synthesis, physical design, multi-modal feature extraction and alignment of circuits.
arXiv Detail & Related papers (2023-12-28T15:09:14Z) - Instructed Language Models with Retrievers Are Powerful Entity Linkers [87.16283281290053]
Instructed Generative Entity Linker (INSGENEL) is the first approach that enables casual language models to perform entity linking over knowledge bases.
INSGENEL outperforms previous generative alternatives with +6.8 F1 points gain on average.
arXiv Detail & Related papers (2023-11-06T16:38:51Z) - Encoder-Decoder Models Can Benefit from Pre-trained Masked Language
Models in Grammatical Error Correction [54.569707226277735]
Previous methods have potential drawbacks when applied to an EncDec model.
Our proposed method fine-tune a corpus and then use the output fine-tuned as additional features in the GEC model.
The best-performing model state-of-the-art performances on the BEA 2019 and CoNLL-2014 benchmarks.
arXiv Detail & Related papers (2020-05-03T04:49:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.