DDPG-Driven Deep-Unfolding with Adaptive Depth for Channel Estimation
with Sparse Bayesian Learning
- URL: http://arxiv.org/abs/2201.08477v3
- Date: Wed, 19 Apr 2023 03:36:19 GMT
- Title: DDPG-Driven Deep-Unfolding with Adaptive Depth for Channel Estimation
with Sparse Bayesian Learning
- Authors: Qiyu Hu, Shuhan Shi, Yunlong Cai and Guanding Yu
- Abstract summary: We first develop a framework of deep deterministic policy gradient (DDPG)-driven deep-unfolding with adaptive depth for different inputs.
Specifically, the framework is employed to deal with the channel estimation problem in massive multiple-input multiple-output systems.
- Score: 23.158142411929322
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep-unfolding neural networks (NNs) have received great attention since they
achieve satisfactory performance with relatively low complexity. Typically,
these deep-unfolding NNs are restricted to a fixed-depth for all inputs.
However, the optimal number of layers required for convergence changes with
different inputs. In this paper, we first develop a framework of deep
deterministic policy gradient (DDPG)-driven deep-unfolding with adaptive depth
for different inputs, where the trainable parameters of deep-unfolding NN are
learned by DDPG, rather than updated by the stochastic gradient descent
algorithm directly. Specifically, the optimization variables, trainable
parameters, and architecture of deep-unfolding NN are designed as the state,
action, and state transition of DDPG, respectively. Then, this framework is
employed to deal with the channel estimation problem in massive multiple-input
multiple-output systems. Specifically, first of all we formulate the channel
estimation problem with an off-grid basis and develop a sparse Bayesian
learning (SBL)-based algorithm to solve it. Secondly, the SBL-based algorithm
is unfolded into a layer-wise structure with a set of introduced trainable
parameters. Thirdly, the proposed DDPG-driven deep-unfolding framework is
employed to solve this channel estimation problem based on the unfolded
structure of the SBL-based algorithm. To realize adaptive depth, we design the
halting score to indicate when to stop, which is a function of the channel
reconstruction error. Furthermore, the proposed framework is extended to
realize the adaptive depth of the general deep neural networks (DNNs).
Simulation results show that the proposed algorithm outperforms the
conventional optimization algorithms and DNNs with fixed depth with much
reduced number of layers.
Related papers
- Broad Critic Deep Actor Reinforcement Learning for Continuous Control [5.440090782797941]
A novel hybrid architecture for actor-critic reinforcement learning (RL) algorithms is introduced.
The proposed architecture integrates the broad learning system (BLS) with deep neural networks (DNNs)
The effectiveness of the proposed algorithm is evaluated by applying it to two classic continuous control tasks.
arXiv Detail & Related papers (2024-11-24T12:24:46Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Unsupervised Learning of Initialization in Deep Neural Networks via
Maximum Mean Discrepancy [74.34895342081407]
We propose an unsupervised algorithm to find good initialization for input data.
We first notice that each parameter configuration in the parameter space corresponds to one particular downstream task of d-way classification.
We then conjecture that the success of learning is directly related to how diverse downstream tasks are in the vicinity of the initial parameters.
arXiv Detail & Related papers (2023-02-08T23:23:28Z) - Neural Basis Functions for Accelerating Solutions to High Mach Euler
Equations [63.8376359764052]
We propose an approach to solving partial differential equations (PDEs) using a set of neural networks.
We regress a set of neural networks onto a reduced order Proper Orthogonal Decomposition (POD) basis.
These networks are then used in combination with a branch network that ingests the parameters of the prescribed PDE to compute a reduced order approximation to the PDE.
arXiv Detail & Related papers (2022-08-02T18:27:13Z) - Adaptive Self-supervision Algorithms for Physics-informed Neural
Networks [59.822151945132525]
Physics-informed neural networks (PINNs) incorporate physical knowledge from the problem domain as a soft constraint on the loss function.
We study the impact of the location of the collocation points on the trainability of these models.
We propose a novel adaptive collocation scheme which progressively allocates more collocation points to areas where the model is making higher errors.
arXiv Detail & Related papers (2022-07-08T18:17:06Z) - On the Turnpike to Design of Deep Neural Nets: Explicit Depth Bounds [0.0]
This paper attempts a quantifiable answer to the question of how many layers should be considered in a Deep Neural Networks (DNN)
The underlying assumption is that the number of neurons per layer -- i.e., the width of the DNN -- is kept constant.
We prove explicit bounds on the required depths of DNNs based on reachability of assumptions and a dissipativity-inducing choice of the regularization terms in the training problem.
arXiv Detail & Related papers (2021-01-08T13:23:37Z) - Solving Sparse Linear Inverse Problems in Communication Systems: A Deep
Learning Approach With Adaptive Depth [51.40441097625201]
We propose an end-to-end trainable deep learning architecture for sparse signal recovery problems.
The proposed method learns how many layers to execute to emit an output, and the network depth is dynamically adjusted for each task in the inference phase.
arXiv Detail & Related papers (2020-10-29T06:32:53Z) - AutoPruning for Deep Neural Network with Dynamic Channel Masking [28.018077874687343]
We propose a learning based auto pruning algorithm for deep neural network.
A two objectives' problem that aims for the the weights and the best channels for each layer is first formulated.
An alternative optimization approach is then proposed to derive the optimal channel numbers and weights simultaneously.
arXiv Detail & Related papers (2020-10-22T20:12:46Z) - Non-Local Spatial Propagation Network for Depth Completion [82.60915972250706]
We propose a robust and efficient end-to-end non-local spatial propagation network for depth completion.
The proposed network takes RGB and sparse depth images as inputs and estimates non-local neighbors and their affinities of each pixel.
We show that the proposed algorithm is superior to conventional algorithms in terms of depth completion accuracy and robustness to the mixed-depth problem.
arXiv Detail & Related papers (2020-07-20T12:26:51Z) - Improving the Backpropagation Algorithm with Consequentialism Weight
Updates over Mini-Batches [0.40611352512781856]
We show that it is possible to consider a multi-layer neural network as a stack of adaptive filters.
We introduce a better algorithm by predicting then emending the adverse consequences of the actions that take place in BP even before they happen.
Our experiments show the usefulness of our algorithm in the training of deep neural networks.
arXiv Detail & Related papers (2020-03-11T08:45:36Z) - Semi-Implicit Back Propagation [1.5533842336139065]
We propose a semi-implicit back propagation method for neural network training.
The difference on the neurons are propagated in a backward fashion and the parameters are updated with proximal mapping.
Experiments on both MNIST and CIFAR-10 demonstrate that the proposed algorithm leads to better performance in terms of both loss decreasing and training/validation accuracy.
arXiv Detail & Related papers (2020-02-10T03:26:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.