DeepSeek-Inspired Exploration of RL-based LLMs and Synergy with Wireless Networks: A Survey
- URL: http://arxiv.org/abs/2503.09956v4
- Date: Mon, 20 Oct 2025 06:44:00 GMT
- Title: DeepSeek-Inspired Exploration of RL-based LLMs and Synergy with Wireless Networks: A Survey
- Authors: Yu Qiao, Phuong-Nam Tran, Ji Su Yoon, Loc X. Nguyen, Eui-Nam Huh, Dusit Niyato, Choong Seon Hong,
- Abstract summary: Reinforcement learning (RL)-based large language models (LLMs) have attracted widespread attention for their capabilities in multimodal data understanding.<n>The open-source DeepSeek models are famous for their innovative designs, such as large-scale pure RL and cost-efficient training.<n>This survey presents a comprehensive exploration of RL-based LLMs in the context of wireless networks.
- Score: 68.74626395093496
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning (RL)-based large language models (LLMs), such as ChatGPT, DeepSeek, and Grok-3, have attracted widespread attention for their remarkable capabilities in multimodal data understanding. Meanwhile, the rapid expansion of information services has led to a growing demand for AI-enabled wireless networks. The open-source DeepSeek models are famous for their innovative designs, such as large-scale pure RL and cost-efficient training, which make them well-suited for practical deployment in wireless networks. By integrating DeepSeek-style LLMs with wireless infrastructures, a synergistic opportunity arises: the DeepSeek-style LLMs enhance network optimization with strong reasoning and decision-making abilities, while wireless infrastructure enables the broad deployment of these models. Motivated by this convergence, this survey presents a comprehensive DeepSeek-inspired exploration of RL-based LLMs in the context of wireless networks. We begin by reviewing key techniques behind network optimization to establish a foundation for understanding DeepSeek-style LLM integration. Next, we examine recent advancements in RL-based LLMs, using DeepSeek models as a representative example. Building on this, we explore the synergy between the two domains, highlighting motivations, challenges, and potential solutions. Finally, we highlight emerging directions for integrating LLMs with wireless networks, such as quantum, on-device, and neural-symbolic LLM models, as well as embodied AI agents. Overall, this survey offers a comprehensive examination of the interplay between DeepSeek-style LLMs and wireless networks, demonstrating how these domains can mutually enhance each other to drive innovation.
Related papers
- Large Language Model (LLM)-enabled Reinforcement Learning for Wireless Network Optimization [79.27012080083603]
Large language models (LLMs) offer promising tools to enhance reinforcement learning in wireless networks.<n>We propose an LLM-assisted state representation and semantic extraction to enhance the multi-agent reinforcement learning framework.
arXiv Detail & Related papers (2026-01-15T01:42:39Z) - Sensing and Understanding the World over Air: A Large Multimodal Model for Mobile Networks [59.23869884913339]
Wireless-native multi-modal large models (WMLMs) can sense and understand the physical world through multi-modal data.<n>We constructed a GPT-style WMLM model and trained it on a real-world large-scale dataset, leveraging wireless signals as an anchor modality for contrastive learning.
arXiv Detail & Related papers (2025-11-17T07:33:46Z) - WebSailor-V2: Bridging the Chasm to Proprietary Agents via Synthetic Data and Scalable Reinforcement Learning [73.91893534088798]
WebSailor is a complete post-training methodology designed to instill this crucial capability.<n>Our approach involves generating novel, high-uncertainty tasks through structured sampling and information obfuscation.<n>WebSailor significantly outperforms all open-source agents in complex information-seeking tasks.
arXiv Detail & Related papers (2025-09-16T17:57:03Z) - Large Language Models for Wireless Communications: From Adaptation to Autonomy [47.40285060307752]
Large language models (LLMs) offer unprecedented capabilities in reasoning, generalization, and zero-shot learning.<n>This article explores the role of LLMs in transforming wireless systems across three key directions.
arXiv Detail & Related papers (2025-07-29T06:21:10Z) - Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities [62.05713042908654]
This paper provides a review of advances in Large Language Models (LLMs) alignment through the lens of inverse reinforcement learning (IRL)<n>We highlight the necessity of constructing neural reward models from human data and discuss the formal and practical implications of this paradigm shift.
arXiv Detail & Related papers (2025-07-17T14:22:24Z) - Prompt-Tuned LLM-Augmented DRL for Dynamic O-RAN Network Slicing [5.62872273155603]
Large Language Models (LLMs) structure unorganized network feedback into meaningful latent representations.<n>In O-RAN slicing, concepts like SNR, power levels and throughput are semantically related.<n>We introduce a contextualization-based adaptation method that integrates learnable prompts into an LLM-augmented DRL framework.
arXiv Detail & Related papers (2025-05-31T14:12:56Z) - LLM-Guided Open RAN: Empowering Hierarchical RAN Intelligent Control [56.94324843095396]
We propose the empowered hierarchical RIC (LLM-hRIC) framework to improve the collaboration between RICs.
This framework integrates LLMs with reinforcement learning (RL) for efficient network resource management.
We evaluate the LLM-hRIC framework in an integrated access and backhaul (IAB) network setting.
arXiv Detail & Related papers (2025-04-25T04:18:23Z) - Offline and Distributional Reinforcement Learning for Wireless Communications [5.771885923067511]
Traditional online reinforcement learning (RL) and deep RL methods face limitations in real-time wireless networks.
We focus on offline and distributional RL, two advanced RL techniques that can overcome these challenges.
We introduce a novel framework that combines offline and distributional RL for wireless communication applications.
arXiv Detail & Related papers (2025-04-04T09:24:39Z) - A Survey on Large Language Models for Communication, Network, and Service Management: Application Insights, Challenges, and Future Directions [37.427638898804055]
Large Language Models (LLMs) have received tremendous attention due to their unparalleled capabilities in various Natural Language Processing (NLP) tasks.<n>This survey investigates the integration of LLMs across different communication network domains, including mobile networks and related technologies, vehicular networks, cloud-based networks, and fog/edge-based networks.
arXiv Detail & Related papers (2024-12-16T20:01:36Z) - NetOrchLLM: Mastering Wireless Network Orchestration with Large Language Models [11.015852090523229]
Large language models (LLMs) have revolutionized various domains by leveraging their sophisticated natural language understanding capabilities.<n>This paper presents NETORCHLLM, a wireless NETwork ORCHestrator LLM framework that seamlessly orchestrates diverse wireless-specific models.<n>A comprehensive framework is introduced, demonstrating the practical viability of our approach.
arXiv Detail & Related papers (2024-12-13T12:48:15Z) - Generative AI-in-the-loop: Integrating LLMs and GPTs into the Next Generation Networks [11.509880721677156]
Large language models (LLMs) have recently emerged, demonstrating near-human-level performance in cognitive tasks.
We propose the concept of "generative AI-in-the-loop"
We believe that combining LLMs and ML models allows both to leverage their respective capabilities and achieve better results than either model alone.
arXiv Detail & Related papers (2024-06-06T17:25:07Z) - WirelessLLM: Empowering Large Language Models Towards Wireless Intelligence [16.722524706176767]
Large Language Models (LLMs) have sparked interest in their potential to revolutionize wireless communication systems.
Existing studies on LLMs for wireless systems are limited to a direct application for telecom language understanding.
This paper proposes WirelessLLM, a comprehensive framework for adapting and enhancing LLMs to address the unique challenges and requirements of wireless communication networks.
arXiv Detail & Related papers (2024-05-27T11:18:25Z) - Large Language Models (LLMs) Assisted Wireless Network Deployment in Urban Settings [0.21847754147782888]
Large Language Models (LLMs) have revolutionized language understanding and human-like text generation.
This paper explores new techniques to harness the power of LLMs for 6G (6th Generation) wireless communication technologies.
We introduce a novel Reinforcement Learning (RL) based framework that leverages LLMs for network deployment in wireless communications.
arXiv Detail & Related papers (2024-05-22T05:19:51Z) - Personalized Wireless Federated Learning for Large Language Models [75.22457544349668]
Large language models (LLMs) have driven profound transformations in wireless networks.<n>Within wireless environments, the training of LLMs faces significant challenges related to security and privacy.<n>This paper presents a systematic analysis of the training stages of LLMs in wireless networks, including pre-training, instruction tuning, and alignment tuning.
arXiv Detail & Related papers (2024-04-20T02:30:21Z) - Large Multi-Modal Models (LMMs) as Universal Foundation Models for
AI-Native Wireless Systems [57.41621687431203]
Large language models (LLMs) and foundation models have been recently touted as a game-changer for 6G systems.
This paper presents a comprehensive vision on how to design universal foundation models tailored towards the deployment of artificial intelligence (AI)-native networks.
arXiv Detail & Related papers (2024-01-30T00:21:41Z) - The Role of Federated Learning in a Wireless World with Foundation Models [59.8129893837421]
Foundation models (FMs) are general-purpose artificial intelligence (AI) models that have recently enabled multiple brand-new generative AI applications.
Currently, the exploration of the interplay between FMs and federated learning (FL) is still in its nascent stage.
This article explores the extent to which FMs are suitable for FL over wireless networks, including a broad overview of research challenges and opportunities.
arXiv Detail & Related papers (2023-10-06T04:13:10Z) - Federated Learning over Wireless IoT Networks with Optimized
Communication and Resources [98.18365881575805]
Federated learning (FL) as a paradigm of collaborative learning techniques has obtained increasing research attention.
It is of interest to investigate fast responding and accurate FL schemes over wireless systems.
We show that the proposed communication-efficient federated learning framework converges at a strong linear rate.
arXiv Detail & Related papers (2021-10-22T13:25:57Z) - Distributed Learning in Wireless Networks: Recent Progress and Future
Challenges [170.35951727508225]
Next-generation wireless networks will enable many machine learning (ML) tools and applications to analyze various types of data collected by edge devices.
Distributed learning and inference techniques have been proposed as a means to enable edge devices to collaboratively train ML models without raw data exchanges.
This paper provides a comprehensive study of how distributed learning can be efficiently and effectively deployed over wireless edge networks.
arXiv Detail & Related papers (2021-04-05T20:57:56Z) - Intelligent Reflecting Surface Aided Wireless Communications: A Tutorial [64.77665786141166]
Intelligent reflecting surface (IRS) is an enabling technology to engineer the radio signal prorogation in wireless networks.
IRS is capable of dynamically altering wireless channels to enhance the communication performance.
Despite its great potential, IRS faces new challenges to be efficiently integrated into wireless networks.
arXiv Detail & Related papers (2020-07-06T13:59:09Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.