An Iterative LLM Framework for SIBT utilizing RAG-based Adaptive Weight Optimization
- URL: http://arxiv.org/abs/2509.08407v1
- Date: Wed, 10 Sep 2025 08:54:16 GMT
- Title: An Iterative LLM Framework for SIBT utilizing RAG-based Adaptive Weight Optimization
- Authors: Zhuo Xiao, Qinglong Yao, Jingjing Wang, Fugen Zhou, Bo Liu, Haitao Sun, Zhe Ji, Yuliang Jiang, Junjie Wang, Qiuwen Wu,
- Abstract summary: This study proposes an adaptive weight optimization framework for SIBT planning, driven by large language models (LLMs)<n>A clinical knowledge base, constructed and queried via retrieval-augmented generation (RAG), enhances the model's domain-specific reasoning.<n>The proposed method was validated on 23 patient cases, showing that the LLM-assisted approach produces plans that are comparable to or exceeding clinically approved and fixed-weight plans.
- Score: 11.168299220031662
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Seed implant brachytherapy (SIBT) is an effective cancer treatment modality; however, clinical planning often relies on manual adjustment of objective function weights, leading to inefficiencies and suboptimal results. This study proposes an adaptive weight optimization framework for SIBT planning, driven by large language models (LLMs). A locally deployed DeepSeek-R1 LLM is integrated with an automatic planning algorithm in an iterative loop. Starting with fixed weights, the LLM evaluates plan quality and recommends new weights in the next iteration. This process continues until convergence criteria are met, after which the LLM conducts a comprehensive evaluation to identify the optimal plan. A clinical knowledge base, constructed and queried via retrieval-augmented generation (RAG), enhances the model's domain-specific reasoning. The proposed method was validated on 23 patient cases, showing that the LLM-assisted approach produces plans that are comparable to or exceeding clinically approved and fixed-weight plans, in terms of dose homogeneity for the clinical target volume (CTV) and sparing of organs at risk (OARs). The study demonstrates the potential use of LLMs in SIBT planning automation.
Related papers
- A Federated and Parameter-Efficient Framework for Large Language Model Training in Medicine [59.78991974851707]
Large language models (LLMs) have demonstrated strong performance on medical benchmarks, including question answering and diagnosis.<n>Most medical LLMs are trained on data from a single institution, which faces limitations in generalizability and safety in heterogeneous systems.<n>We introduce the model-agnostic and parameter-efficient federated learning framework for adapting LLMs to medical applications.
arXiv Detail & Related papers (2026-01-29T18:48:21Z) - Zero-Shot Large Language Model Agents for Fully Automated Radiotherapy Treatment Planning [14.814676057920067]
A large language model (LLM)-based agent navigates inverse treatment planning for intensity-modulated radiation therapy (IMRT)<n>The agent's decision-making process is informed by current observations and previous optimization attempts and evaluations.<n>This study demonstrates the feasibility of a zero-shot, LLM-driven workflow for automated IMRT treatment planning in a commercial TPS.
arXiv Detail & Related papers (2025-10-12T19:21:21Z) - Stabilizing Policy Gradients for Sample-Efficient Reinforcement Learning in LLM Reasoning [77.92320830700797]
Reinforcement Learning has played a central role in enabling reasoning capabilities of Large Language Models.<n>We propose a tractable computational framework that tracks and leverages curvature information during policy updates.<n>The algorithm, Curvature-Aware Policy Optimization (CAPO), identifies samples that contribute to unstable updates and masks them out.
arXiv Detail & Related papers (2025-10-01T12:29:32Z) - Learn to optimize for automatic proton PBS treatment planning for H&N cancers [3.384581246727367]
We propose a data-driven inverse and integrate it into a PPO-based automatic treatment planning framework.<n>The PPO framework functions as an outer-loop virtual planner, autonomously adjusting objective parameters through a policy network.<n>Compared with L-BFGSB, our L2O-based inverse improves the effectiveness and efficiency by 22.97% and 36.41%, respectively.
arXiv Detail & Related papers (2025-08-14T21:50:31Z) - Automated Treatment Planning for Interstitial HDR Brachytherapy for Locally Advanced Cervical Cancer using Deep Reinforcement Learning [3.9838929530763076]
The objective of this study is to develop a fully automated HDR brachytherapy planning framework.<n>We propose a hierarchical two-stage autoplanning framework.<n>For the unseen test patients, the RL-based automated planning method achieved an average score of 93.89%, outperforming the clinical plans which averaged 91.86%.
arXiv Detail & Related papers (2025-06-13T17:07:30Z) - Patient-Specific Deep Reinforcement Learning for Automatic Replanning in Head-and-Neck Cancer Proton Therapy [8.677300387603356]
Anatomical changes during intensity-modulated proton therapy (IMPT) for head-and-neck cancer (HNC) can shift Bragg peaks, risking tumor underdosing and organ-at-risk overdosing.<n>We propose a patient-specific deep reinforcement learning framework for automated IMPT replanning.
arXiv Detail & Related papers (2025-06-11T18:00:06Z) - Autonomous Radiotherapy Treatment Planning Using DOLA: A Privacy-Preserving, LLM-Based Optimization Agent [2.1986172572830096]
Dose Optimization Language Agent (DOLA) is an autonomous large language model (LLM)-based agent designed for optimizing radiotherapy treatment plans.<n>DOLA integrates the LLaMa3.1 LLM directly with a commercial treatment planning system.<n> operating entirely within secure local infrastructure.
arXiv Detail & Related papers (2025-03-21T22:01:19Z) - Attribute Controlled Fine-tuning for Large Language Models: A Case Study on Detoxification [76.14641982122696]
We propose a constraint learning schema for fine-tuning Large Language Models (LLMs) with attribute control.
We show that our approach leads to an LLM that produces fewer inappropriate responses while achieving competitive performance on benchmarks and a toxicity detection task.
arXiv Detail & Related papers (2024-10-07T23:38:58Z) - Automating proton PBS treatment planning for head and neck cancers using policy gradient-based deep reinforcement learning [0.7519872646378836]
We propose an automatic treatment planning model using the proximal policy optimization (PPO) algorithm and a dose distribution-based reward function.
A set of empirical rules is used to create auxiliary planning structures from target volumes and organs-at-risk.
A decision-making policy network trained using PPO is developed to iteratively adjust the involved planning objective parameters in a continuous action space.
arXiv Detail & Related papers (2024-09-17T22:01:56Z) - Exploring and Benchmarking the Planning Capabilities of Large Language Models [57.23454975238014]
This work lays the foundations for improving planning capabilities of large language models (LLMs)
We construct a comprehensive benchmark suite encompassing both classical planning benchmarks and natural language scenarios.
We investigate the use of many-shot in-context learning to enhance LLM planning, exploring the relationship between increased context length and improved planning performance.
arXiv Detail & Related papers (2024-06-18T22:57:06Z) - End-to-End Breast Cancer Radiotherapy Planning via LMMs with Consistency Embedding [47.360760580820966]
We present RO-LMM, a comprehensive large multimodal model (LMM) tailored for the field of radiation oncology.<n>This model effectively manages a series of tasks within the clinical workflow, including clinical context summarization, radiation treatment plan suggestion, and plan-guided target volume segmentation.<n>We present a novel Consistency Embedding Fine-Tuning (CEFTune) technique, which boosts LMM's robustness to noisy inputs while preserving the consistency of handling clean inputs.
arXiv Detail & Related papers (2023-11-27T14:49:06Z) - Resource Planning for Hospitals Under Special Consideration of the
COVID-19 Pandemic: Optimization and Sensitivity Analysis [87.31348761201716]
Crises like the COVID-19 pandemic pose a serious challenge to health-care institutions.
BaBSim.Hospital is a tool for capacity planning based on discrete event simulation.
We aim to investigate and optimize these parameters to improve BaBSim.Hospital.
arXiv Detail & Related papers (2021-05-16T12:38:35Z) - A feasibility study of a hyperparameter tuning approach to automated
inverse planning in radiotherapy [68.8204255655161]
The purpose of this study is to automate the inverse planning process to reduce active planning time while maintaining plan quality.
We investigated the impact of the choice of dose parameters, random and Bayesian search methods, and utility function form on planning time and plan quality.
Using 100 samples was found to produce satisfactory plan quality, and the average planning time was 2.3 hours.
arXiv Detail & Related papers (2021-05-14T18:37:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.