Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork
- URL: http://arxiv.org/abs/2101.04348v1
- Date: Tue, 12 Jan 2021 08:36:23 GMT
- Title: Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork
- Authors: Chang-Jen Wang, Chao-Kai Wen, Shang-Ho (Lawrence) Tsai, Shi Jin,
Geoffrey Ye Li
- Abstract summary: Phase retrieval is an important component in modern computational imaging systems.
Recent advances in deep learning have opened up a new possibility for robust and fast PR.
We develop a novel framework for deep unfolding to overcome the existing limitations.
- Score: 73.94896986868146
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Phase retrieval (PR) is an important component in modern computational
imaging systems. Many algorithms have been developed over the past half
century. Recent advances in deep learning have opened up a new possibility for
robust and fast PR. An emerging technique, called deep unfolding, provides a
systematic connection between conventional model-based iterative algorithms and
modern data-based deep learning. Unfolded algorithms, powered by data learning,
have shown remarkable performance and convergence speed improvement over the
original algorithms. Despite their potential, most existing unfolded algorithms
are strictly confined to a fixed number of iterations when employing
layer-dependent parameters. In this study, we develop a novel framework for
deep unfolding to overcome the existing limitations. Even if our framework can
be widely applied to general inverse problems, we take PR as an example in the
paper. Our development is based on an unfolded generalized expectation
consistent signal recovery (GEC-SR) algorithm, wherein damping factors are left
for data-driven learning. In particular, we introduce a hypernetwork to
generate the damping factors for GEC-SR. Instead of directly learning a set of
optimal damping factors, the hypernetwork learns how to generate the optimal
damping factors according to the clinical settings, thus ensuring its
adaptivity to different scenarios. To make the hypernetwork work adapt to
varying layer numbers, we use a recurrent architecture to develop a dynamic
hypernetwork, which generates a damping factor that can vary online across
layers. We also exploit a self-attention mechanism to enhance the robustness of
the hypernetwork. Extensive experiments show that the proposed algorithm
outperforms existing ones in convergence speed and accuracy, and still works
well under very harsh settings, that many classical PR algorithms unstable or
even fail.
Related papers
- Adaptive Anomaly Detection in Network Flows with Low-Rank Tensor Decompositions and Deep Unrolling [9.20186865054847]
Anomaly detection (AD) is increasingly recognized as a key component for ensuring the resilience of future communication systems.
This work considers AD in network flows using incomplete measurements.
We propose a novel block-successive convex approximation algorithm based on a regularized model-fitting objective.
Inspired by Bayesian approaches, we extend the model architecture to perform online adaptation to per-flow and per-time-step statistics.
arXiv Detail & Related papers (2024-09-17T19:59:57Z) - Forward-Forward Algorithm for Hyperspectral Image Classification: A
Preliminary Study [0.0]
Forward-forward algorithm (FFA) computes local goodness functions to optimize network parameters.
This study investigates the application of FFA for hyperspectral image classification.
arXiv Detail & Related papers (2023-07-01T05:39:28Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Biologically Plausible Learning on Neuromorphic Hardware Architectures [27.138481022472]
Neuromorphic computing is an emerging paradigm that confronts this imbalance by computations directly in analog memories.
This work is the first to compare the impact of different learning algorithms on Compute-In-Memory-based hardware and vice versa.
arXiv Detail & Related papers (2022-12-29T15:10:59Z) - Analytically Tractable Inference in Deep Neural Networks [0.0]
Tractable Approximate Inference (TAGI) algorithm was shown to be a viable and scalable alternative to backpropagation for shallow fully-connected neural networks.
We are demonstrating how TAGI matches or exceeds the performance of backpropagation, for training classic deep neural network architectures.
arXiv Detail & Related papers (2021-03-09T14:51:34Z) - Solving Sparse Linear Inverse Problems in Communication Systems: A Deep
Learning Approach With Adaptive Depth [51.40441097625201]
We propose an end-to-end trainable deep learning architecture for sparse signal recovery problems.
The proposed method learns how many layers to execute to emit an output, and the network depth is dynamically adjusted for each task in the inference phase.
arXiv Detail & Related papers (2020-10-29T06:32:53Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.