Deep Transfer Learning Based Downlink Channel Prediction for FDD Massive
MIMO Systems
- URL: http://arxiv.org/abs/1912.12265v4
- Date: Mon, 7 Sep 2020 11:43:30 GMT
- Title: Deep Transfer Learning Based Downlink Channel Prediction for FDD Massive
MIMO Systems
- Authors: Yuwen Yang, Feifei Gao, Zhimeng Zhong, Bo Ai and Ahmed Alkhateeb
- Abstract summary: We formulate the downlink channel prediction as a deep transfer learning (DTL) problem.
Specifically, we develop the direct-transfer algorithm based on the fully-connected neural network architecture.
To further improve the transfer efficiency, we propose the meta-learning algorithm that trains the network by alternating inner-task and across-task updates.
- Score: 43.63380272164857
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial intelligence (AI) based downlink channel state information (CSI)
prediction for frequency division duplexing (FDD) massive multiple-input
multiple-output (MIMO) systems has attracted growing attention recently.
However, existing works focus on the downlink CSI prediction for the users
under a given environment and is hard to adapt to users in new environment
especially when labeled data is limited. To address this issue, we formulate
the downlink channel prediction as a deep transfer learning (DTL) problem,
where each learning task aims to predict the downlink CSI from the uplink CSI
for one single environment. Specifically, we develop the direct-transfer
algorithm based on the fully-connected neural network architecture, where the
network is trained on the data from all previous environments in the manner of
classical deep learning and is then fine-tuned for new environments. To further
improve the transfer efficiency, we propose the meta-learning algorithm that
trains the network by alternating inner-task and across-task updates and then
adapts to a new environment with a small number of labeled data. Simulation
results show that the direct-transfer algorithm achieves better performance
than the deep learning algorithm, which implies that the transfer learning
benefits the downlink channel prediction in new environments. Moreover, the
meta-learning algorithm significantly outperforms the direct-transfer algorithm
in terms of both prediction accuracy and stability, which validates its
effectiveness and superiority.
Related papers
- Efficient Model Adaptation for Continual Learning at the Edge [15.334881190102895]
Most machine learning (ML) systems assume stationary and matching data distributions during training and deployment.
Data distributions often shift over time due to changes in environmental factors, sensor characteristics, and task-of-interest.
This paper presents theAdaptor-Reconfigurator (EAR) framework for efficient continual learning under domain shifts.
arXiv Detail & Related papers (2023-08-03T23:55:17Z) - Transfer Learning Based Efficient Traffic Prediction with Limited
Training Data [3.689539481706835]
Efficient prediction of internet traffic is an essential part of Self Organizing Network (SON) for ensuring proactive management.
Deep sequence model in network traffic prediction with limited training data has not been studied extensively in the current works.
We investigated and evaluated the performance of the deep transfer learning technique in traffic prediction with inadequate historical data.
arXiv Detail & Related papers (2022-05-09T14:44:39Z) - Predictive coding, precision and natural gradients [2.1601966913620325]
We show that hierarchical predictive coding networks with learnable precision are able to solve various supervised and unsupervised learning tasks.
When applied to unsupervised auto-encoding of image inputs, the deterministic network produces hierarchically organized and disentangled embeddings.
arXiv Detail & Related papers (2021-11-12T21:05:03Z) - Learning to Continuously Optimize Wireless Resource in a Dynamic
Environment: A Bilevel Optimization Perspective [52.497514255040514]
This work develops a new approach that enables data-driven methods to continuously learn and optimize resource allocation strategies in a dynamic environment.
We propose to build the notion of continual learning into wireless system design, so that the learning model can incrementally adapt to the new episodes.
Our design is based on a novel bilevel optimization formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2021-05-03T07:23:39Z) - A Meta-Learning Approach to the Optimal Power Flow Problem Under
Topology Reconfigurations [69.73803123972297]
We propose a DNN-based OPF predictor that is trained using a meta-learning (MTL) approach.
The developed OPF-predictor is validated through simulations using benchmark IEEE bus systems.
arXiv Detail & Related papers (2020-12-21T17:39:51Z) - Learning to Continuously Optimize Wireless Resource In Episodically
Dynamic Environment [55.91291559442884]
This work develops a methodology that enables data-driven methods to continuously learn and optimize in a dynamic environment.
We propose to build the notion of continual learning into the modeling process of learning wireless systems.
Our design is based on a novel min-max formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2020-11-16T08:24:34Z) - Regularizing Deep Networks with Semantic Data Augmentation [44.53483945155832]
We propose a novel semantic data augmentation algorithm to complement traditional approaches.
The proposed method is inspired by the intriguing property that deep networks are effective in learning linearized features.
We show that the proposed implicit semantic data augmentation (ISDA) algorithm amounts to minimizing a novel robust CE loss.
arXiv Detail & Related papers (2020-07-21T00:32:44Z) - Deep Transfer Learning with Ridge Regression [7.843067454030999]
Deep models trained with massive amounts of data demonstrate promising generalisation ability on unseen data from relevant domains.
We address this issue by leveraging the low-rank property of learnt feature vectors produced from deep neural networks (DNNs) with the closed-form solution provided in kernel ridge regression (KRR)
Our method is successful on supervised and semi-supervised transfer learning tasks.
arXiv Detail & Related papers (2020-06-11T20:21:35Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.