Denoising Noisy Neural Networks: A Bayesian Approach with Compensation
- URL: http://arxiv.org/abs/2105.10699v1
- Date: Sat, 22 May 2021 11:51:20 GMT
- Title: Denoising Noisy Neural Networks: A Bayesian Approach with Compensation
- Authors: Yulin Shao and Soung Chang Liew and Deniz Gunduz
- Abstract summary: Noisy neural networks (NoisyNNs) refer to the inference and training of NNs in the presence of noise.
This paper studies how to estimate the uncontaminated NN weights from their noisy observations or manifestations.
- Score: 36.39188653838991
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Noisy neural networks (NoisyNNs) refer to the inference and training of NNs
in the presence of noise. Noise is inherent in most communication and storage
systems; hence, NoisyNNs emerge in many new applications, including federated
edge learning, where wireless devices collaboratively train a NN over a noisy
wireless channel, or when NNs are implemented/stored in an analog storage
medium. This paper studies a fundamental problem of NoisyNNs: how to estimate
the uncontaminated NN weights from their noisy observations or manifestations.
Whereas all prior works relied on the maximum likelihood (ML) estimation to
maximize the likelihood function of the estimated NN weights, this paper
demonstrates that the ML estimator is in general suboptimal. To overcome the
suboptimality of the conventional ML estimator, we put forth an
$\text{MMSE}_{pb}$ estimator to minimize a compensated mean squared error (MSE)
with a population compensator and a bias compensator. Our approach works well
for NoisyNNs arising in both 1) noisy inference, where noise is introduced only
in the inference phase on the already-trained NN weights; and 2) noisy
training, where noise is introduced over the course of training. Extensive
experiments on the CIFAR-10 and SST-2 datasets with different NN architectures
verify the significant performance gains of the $\text{MMSE}_{pb}$ estimator
over the ML estimator when used to denoise the NoisyNN. For noisy inference,
the average gains are up to $156\%$ for a noisy ResNet34 model and $14.7\%$ for
a noisy BERT model; for noisy training, the average gains are up to $18.1$ dB
for a noisy ResNet18 model.
Related papers
- Neural Network-based Vehicular Channel Estimation Performance: Effect of Noise in the Training Set [0.5530212768657544]
Vehicular communication systems face significant challenges due to high mobility and rapidly changing environments.
To address these challenges, neural network (NN)-based channel estimation methods have been suggested.
This study examines the effectiveness of training NN-based channel estimators on mixed SNR datasets.
arXiv Detail & Related papers (2025-02-05T09:29:01Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Understanding and Mitigating the Label Noise in Pre-training on
Downstream Tasks [91.15120211190519]
This paper aims to understand the nature of noise in pre-training datasets and to mitigate its impact on downstream tasks.
We propose a light-weight black-box tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise.
arXiv Detail & Related papers (2023-09-29T06:18:15Z) - Latent Class-Conditional Noise Model [54.56899309997246]
We introduce a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework.
We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels.
Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples.
arXiv Detail & Related papers (2023-02-19T15:24:37Z) - Impact of L1 Batch Normalization on Analog Noise Resistant Property of
Deep Learning Models [3.520496620951778]
In this work, the use of L1 or TopK BatchNorm type in designing deep neural network (DNN) models with excellent noise-resistant property is proposed.
The resulting model noise-resistant property is tested by injecting additive noise to the model weights and evaluating the new model inference accuracy due to the noise.
The results show that L1 and TopK BatchNorm type has excellent noise-resistant property, and there is no sacrifice in performance due to the change in the BatchNorm type from L2 to L1/TopK BatchNorm type.
arXiv Detail & Related papers (2022-05-07T22:23:21Z) - Robust Learning of Recurrent Neural Networks in Presence of Exogenous
Noise [22.690064709532873]
We propose a tractable robustness analysis for RNN models subject to input noise.
The robustness measure can be estimated efficiently using linearization techniques.
Our proposed methodology significantly improves robustness of recurrent neural networks.
arXiv Detail & Related papers (2021-05-03T16:45:05Z) - Learning based signal detection for MIMO systems with unknown noise
statistics [84.02122699723536]
This paper aims to devise a generalized maximum likelihood (ML) estimator to robustly detect signals with unknown noise statistics.
In practice, there is little or even no statistical knowledge on the system noise, which in many cases is non-Gaussian, impulsive and not analyzable.
Our framework is driven by an unsupervised learning approach, where only the noise samples are required.
arXiv Detail & Related papers (2021-01-21T04:48:15Z) - Deep Networks for Direction-of-Arrival Estimation in Low SNR [89.45026632977456]
We introduce a Convolutional Neural Network (CNN) that is trained from mutli-channel data of the true array manifold matrix.
We train a CNN in the low-SNR regime to predict DoAs across all SNRs.
Our robust solution can be applied in several fields, ranging from wireless array sensors to acoustic microphones or sonars.
arXiv Detail & Related papers (2020-11-17T12:52:18Z) - Robust Processing-In-Memory Neural Networks via Noise-Aware
Normalization [26.270754571140735]
PIM accelerators often suffer from intrinsic noise in the physical components.
We propose a noise-agnostic method to achieve robust neural network performance against any noise setting.
arXiv Detail & Related papers (2020-07-07T06:51:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.