Physics-Based Deep Learning for Fiber-Optic Communication Systems
- URL: http://arxiv.org/abs/2010.14258v1
- Date: Tue, 27 Oct 2020 12:55:23 GMT
- Title: Physics-Based Deep Learning for Fiber-Optic Communication Systems
- Authors: Christian H\"ager and Henry D. Pfister
- Abstract summary: We propose a new machine-learning approach for fiber-optic communication systems governed by the nonlinear Schr"odinger equation (NLSE)
Our main observation is that the popular split-step method (SSM) for numerically solving the NLSE has essentially the same functional form as a deep multi-layer neural network.
We exploit this connection by parameterizing the SSM and viewing the linear steps as general linear functions, similar to the weight matrices in a neural network.
- Score: 10.630021520220653
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a new machine-learning approach for fiber-optic communication
systems whose signal propagation is governed by the nonlinear Schr\"odinger
equation (NLSE). Our main observation is that the popular split-step method
(SSM) for numerically solving the NLSE has essentially the same functional form
as a deep multi-layer neural network; in both cases, one alternates linear
steps and pointwise nonlinearities. We exploit this connection by
parameterizing the SSM and viewing the linear steps as general linear
functions, similar to the weight matrices in a neural network. The resulting
physics-based machine-learning model has several advantages over "black-box"
function approximators. For example, it allows us to examine and interpret the
learned solutions in order to understand why they perform well. As an
application, low-complexity nonlinear equalization is considered, where the
task is to efficiently invert the NLSE. This is commonly referred to as digital
backpropagation (DBP). Rather than employing neural networks, the proposed
algorithm, dubbed learned DBP (LDBP), uses the physics-based model with
trainable filters in each step and its complexity is reduced by progressively
pruning filter taps during gradient descent. Our main finding is that the
filters can be pruned to remarkably short lengths-as few as 3 taps/step-without
sacrificing performance. As a result, the complexity can be reduced by orders
of magnitude in comparison to prior work. By inspecting the filter responses,
an additional theoretical justification for the learned parameter
configurations is provided. Our work illustrates that combining data-driven
optimization with existing domain knowledge can generate new insights into old
communications problems.
Related papers
- Nonlinear Schrödinger Network [0.8249694498830558]
Deep neural networks (DNNs) have achieved exceptional performance across various fields by learning complex nonlinear mappings from large-scale datasets.
To address these issues, hybrid approaches that integrate physics with AI are gaining interest.
This paper introduces a novel physics-based AI model called the "Nonlinear Schr"odinger Network"
arXiv Detail & Related papers (2024-07-19T17:58:00Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - AMS-Net: Adaptive Multiscale Sparse Neural Network with Interpretable
Basis Expansion for Multiphase Flow Problems [8.991619150027267]
We propose an adaptive sparse learning algorithm that can be applied to learn the physical processes and obtain a sparse representation of the solution given a large snapshot space.
The information of the basis functions are incorporated in the loss function, which minimizes the differences between the downscaled reduced order solutions and reference solutions at multiple time steps.
More numerical tests are performed on two-phase multiscale flow problems to show the capability and interpretability of the proposed method on complicated applications.
arXiv Detail & Related papers (2022-07-24T13:12:43Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Physics-informed Neural Network for Nonlinear Dynamics in Fiber Optics [10.335960060544904]
A physics-informed neural network (PINN) that combines deep learning with physics is studied to solve the nonlinear Schr"odinger equation for learning nonlinear dynamics in fiber optics.
PINN is not only an effective partial differential equation solver, but also a prospective technique to advance the scientific computing and automatic modeling in fiber optics.
arXiv Detail & Related papers (2021-09-01T12:19:32Z) - Inverse Problem of Nonlinear Schr\"odinger Equation as Learning of
Convolutional Neural Network [5.676923179244324]
It is shown that one can obtain a relatively accurate estimate of the considered parameters using the proposed method.
It provides a natural framework in inverse problems of partial differential equations with deep learning.
arXiv Detail & Related papers (2021-07-19T02:54:37Z) - A Meta-Learning Approach to the Optimal Power Flow Problem Under
Topology Reconfigurations [69.73803123972297]
We propose a DNN-based OPF predictor that is trained using a meta-learning (MTL) approach.
The developed OPF-predictor is validated through simulations using benchmark IEEE bus systems.
arXiv Detail & Related papers (2020-12-21T17:39:51Z) - Solving Sparse Linear Inverse Problems in Communication Systems: A Deep
Learning Approach With Adaptive Depth [51.40441097625201]
We propose an end-to-end trainable deep learning architecture for sparse signal recovery problems.
The proposed method learns how many layers to execute to emit an output, and the network depth is dynamically adjusted for each task in the inference phase.
arXiv Detail & Related papers (2020-10-29T06:32:53Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - A Neural Network Approach for Online Nonlinear Neyman-Pearson
Classification [3.6144103736375857]
We propose a novel Neyman-Pearson (NP) classifier that is both online and nonlinear as the first time in the literature.
The proposed classifier operates on a binary labeled data stream in an online manner, and maximizes the detection power about a user-specified and controllable false positive rate.
Our algorithm is appropriate for large scale data applications and provides a decent false positive rate controllability with real time processing.
arXiv Detail & Related papers (2020-06-14T20:00:25Z) - Computational optimization of convolutional neural networks using
separated filters architecture [69.73393478582027]
We consider a convolutional neural network transformation that reduces computation complexity and thus speedups neural network processing.
Use of convolutional neural networks (CNN) is the standard approach to image recognition despite the fact they can be too computationally demanding.
arXiv Detail & Related papers (2020-02-18T17:42:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.