LLM-Guided Planning and Summary-Based Scientific Text Simplification: DS@GT at CLEF 2025 SimpleText
- URL: http://arxiv.org/abs/2508.11816v1
- Date: Fri, 15 Aug 2025 21:44:52 GMT
- Title: LLM-Guided Planning and Summary-Based Scientific Text Simplification: DS@GT at CLEF 2025 SimpleText
- Authors: Krishna Chaitanya Marturi, Heba H. Elwazzan,
- Abstract summary: We present our approach for the CLEF 2025 SimpleText Task 1, which addresses both sentence-level and document-level scientific text simplification.<n>For sentence-level simplification, our methodology employs large language models (LLMs) to first generate a structured plan, followed by plan-driven simplification of individual sentences.<n>At the document level, we leverage LLMs to produce concise summaries and subsequently guide the simplification process using these summaries.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present our approach for the CLEF 2025 SimpleText Task 1, which addresses both sentence-level and document-level scientific text simplification. For sentence-level simplification, our methodology employs large language models (LLMs) to first generate a structured plan, followed by plan-driven simplification of individual sentences. At the document level, we leverage LLMs to produce concise summaries and subsequently guide the simplification process using these summaries. This two-stage, LLM-based framework enables more coherent and contextually faithful simplifications of scientific text.
Related papers
- OUNLP at TSAR 2025 Shared Task: Multi-Round Text Simplifier via Code Generation [3.117899569620301]
This paper describes the OUNLP system submitted to the TSAR-2025 Shared Task (Alva-Manchego et al., 2025)<n>It is designed for readability-controlled text simplification using LLM-prompting-based generation.
arXiv Detail & Related papers (2025-11-06T16:16:32Z) - Context-Aware Hierarchical Taxonomy Generation for Scientific Papers via LLM-Guided Multi-Aspect Clustering [59.54662810933882]
Existing taxonomy construction methods, leveraging unsupervised clustering or direct prompting of large language models, often lack coherence and granularity.<n>We propose a novel context-aware hierarchical taxonomy generation framework that integrates LLM-guided multi-aspect encoding with dynamic clustering.
arXiv Detail & Related papers (2025-09-23T15:12:58Z) - Hallucination Detection and Mitigation in Scientific Text Simplification using Ensemble Approaches: DS@GT at CLEF 2025 SimpleText [0.0]
We describe our methodology for the CLEF 2025 SimpleText Task 2, which focuses on detecting and evaluating creative generation and information distortion.<n>Our solution integrates multiple strategies: we construct an ensemble framework that leverages BERT-based classifier, semantic similarity measure, natural language inference model, and large language model.<n>For grounded generation, we employ an LLM-based post-editing system that revises simplifications based on the original input texts.
arXiv Detail & Related papers (2025-08-15T21:57:27Z) - Aligning Sentence Simplification with ESL Learner's Proficiency for Language Acquisition [11.700462697630696]
This study aims to facilitate English as a Second Language learners' language acquisition by simplification.<n>We propose simplifying complex sentences to appropriate levels for learners while also increasing vocabulary coverage of the target level in the simplifications.<n>Our method employs token-level and sentence-level rewards, and iteratively trains the model on its self-generated outputs to guide the model to search for simplification hypotheses that satisfy the target attributes.
arXiv Detail & Related papers (2025-02-17T05:32:56Z) - Progressive Document-level Text Simplification via Large Language Models [19.57555397986868]
Long document-level simplification (DS) is still relatively unexplored.<n>We propose a progressive simplification method (ProgDS) by hierarchically decomposing the task.
arXiv Detail & Related papers (2025-01-07T15:14:37Z) - Simple is not Enough: Document-level Text Simplification using Readability and Coherence [20.613410797137036]
We present the SimDoc system, a simplification model considering simplicity, readability, and discourse aspects, such as coherence.<n>We include multiple objectives during training, considering simplicity, readability, and coherence altogether.<n>We present a comparative analysis in which we evaluate our proposed models in a zero-shot, few-shot, and fine-tuning setting using document-level TS corpora.
arXiv Detail & Related papers (2024-12-24T19:05:21Z) - Integrating Planning into Single-Turn Long-Form Text Generation [66.08871753377055]
We propose to use planning to generate long form content.
Our main novelty lies in a single auxiliary task that does not require multiple rounds of prompting or planning.
Our experiments demonstrate on two datasets from different domains, that LLMs fine-tuned with the auxiliary task generate higher quality documents.
arXiv Detail & Related papers (2024-10-08T17:02:40Z) - DECIDER: A Dual-System Rule-Controllable Decoding Framework for Language Generation [57.07295906718989]
Constrained decoding approaches aim to control the meaning or style of text generated by pre-trained large language (Ms also PLMs) for various tasks at inference time.<n>These methods often guide plausible continuations by greedily and explicitly selecting targets.<n>Inspired by cognitive dual-process theory, we propose a novel decoding framework DECIDER.
arXiv Detail & Related papers (2024-03-04T11:49:08Z) - Teaching the Pre-trained Model to Generate Simple Texts for Text
Simplification [59.625179404482594]
Randomly masking text spans in ordinary texts in the pre-training stage hardly allows models to acquire the ability to generate simple texts.
We propose a new continued pre-training strategy to teach the pre-trained model to generate simple texts.
arXiv Detail & Related papers (2023-05-21T14:03:49Z) - Elaborative Simplification as Implicit Questions Under Discussion [51.17933943734872]
This paper proposes to view elaborative simplification through the lens of the Question Under Discussion (QUD) framework.
We show that explicitly modeling QUD provides essential understanding of elaborative simplification and how the elaborations connect with the rest of the discourse.
arXiv Detail & Related papers (2023-05-17T17:26:16Z) - Context-Aware Document Simplification [3.2880869992413237]
We explore systems that use document context within the simplification process itself.
We achieve state-of-the-art performance on the document simplification task, even when not relying on plan-guidance.
arXiv Detail & Related papers (2023-05-10T16:06:36Z) - Controllable Abstractive Dialogue Summarization with Sketch Supervision [56.59357883827276]
Our model achieves state-of-the-art performance on the largest dialogue summarization corpus SAMSum, with as high as 50.79 in ROUGE-L score.
arXiv Detail & Related papers (2021-05-28T19:05:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.