Learning to Learn to Demodulate with Uncertainty Quantification via
Bayesian Meta-Learning
- URL: http://arxiv.org/abs/2108.00785v1
- Date: Mon, 2 Aug 2021 11:07:46 GMT
- Title: Learning to Learn to Demodulate with Uncertainty Quantification via
Bayesian Meta-Learning
- Authors: Kfir M. Cohen, Sangwoo Park, Osvaldo Simeone, Shlomo Shamai (Shitz)
- Abstract summary: We introduce the use of Bayesian meta-learning via variational inference for the purpose of obtaining well-calibrated few-pilot demodulators.
The resulting Bayesian ensembles offer better calibrated soft decisions, at the computational cost of running multiple instances of the neural network for demodulation.
- Score: 59.014197664747165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Meta-learning, or learning to learn, offers a principled framework for
few-shot learning. It leverages data from multiple related learning tasks to
infer an inductive bias that enables fast adaptation on a new task. The
application of meta-learning was recently proposed for learning how to
demodulate from few pilots. The idea is to use pilots received and stored for
offline use from multiple devices in order to meta-learn an adaptation
procedure with the aim of speeding up online training on new devices. Standard
frequentist learning, which can yield relatively accurate "hard" classification
decisions, is known to be poorly calibrated, particularly in the small-data
regime. Poor calibration implies that the soft scores output by the demodulator
are inaccurate estimates of the true probability of correct demodulation. In
this work, we introduce the use of Bayesian meta-learning via variational
inference for the purpose of obtaining well-calibrated few-pilot demodulators.
In a Bayesian framework, each neural network weight is represented by a
distribution, capturing epistemic uncertainty. Bayesian meta-learning optimizes
over the prior distribution of the weights. The resulting Bayesian ensembles
offer better calibrated soft decisions, at the computational cost of running
multiple instances of the neural network for demodulation. Numerical results
for single-input single-output Rayleigh fading channels with transmitter's
non-linearities are provided that compare symbol error rate and expected
calibration error for both frequentist and Bayesian meta-learning, illustrating
how the latter is both more accurate and better-calibrated.
Related papers
- Gradient-free variational learning with conditional mixture networks [39.827869318925494]
Conditional mixture networks (CMNs) are suitable for fast, gradient-free inference and can solve complex classification tasks.
We validate this approach by training two-layer CMNs on standard benchmarks from the UCI repository.
Our method, CAVI-CMN, achieves competitive and often superior predictive accuracy compared to maximum likelihood estimation (MLE) with backpropagation.
arXiv Detail & Related papers (2024-08-29T10:43:55Z) - Unsupervised Representation Learning to Aid Semi-Supervised Meta
Learning [16.534014215010757]
We propose a one-shot unsupervised meta-learning to learn latent representation of training samples.
A temperature-scaled cross-entropy loss is used in the inner loop of meta-learning to prevent overfitting.
The proposed method is model agnostic and can aid any meta-learning model to improve accuracy.
arXiv Detail & Related papers (2023-10-19T18:25:22Z) - Calibration-Aware Bayesian Learning [37.82259435084825]
This paper proposes an integrated framework, referred to as calibration-aware Bayesian neural networks (CA-BNNs)
It applies both data-dependent or data-independent regularizers while optimizing over a variational distribution as in Bayesian learning.
Numerical results validate the advantages of the proposed approach in terms of expected calibration error (ECE) and reliability diagrams.
arXiv Detail & Related papers (2023-05-12T14:19:15Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Multi-Head Multi-Loss Model Calibration [13.841172927454204]
We introduce a form of simplified ensembling that bypasses the costly training and inference of deep ensembles.
Specifically, each head is trained to minimize a weighted Cross-Entropy loss, but the weights are different among the different branches.
We show that the resulting averaged predictions can achieve excellent calibration without sacrificing accuracy in two challenging datasets.
arXiv Detail & Related papers (2023-03-02T09:32:32Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Distribution Mismatch Correction for Improved Robustness in Deep Neural
Networks [86.42889611784855]
normalization methods increase the vulnerability with respect to noise and input corruptions.
We propose an unsupervised non-parametric distribution correction method that adapts the activation distribution of each layer.
In our experiments, we empirically show that the proposed method effectively reduces the impact of intense image corruptions.
arXiv Detail & Related papers (2021-10-05T11:36:25Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - Robust Cell-Load Learning with a Small Sample Set [35.07023055409166]
Learning of the cell-load in radio access networks (RANs) has to be performed within a short time period.
We propose a learning framework that is robust against uncertainties resulting from the need for learning based on a relatively small training sample set.
arXiv Detail & Related papers (2021-03-21T19:17:01Z) - Learning Neural Network Subspaces [74.44457651546728]
Recent observations have advanced our understanding of the neural network optimization landscape.
With a similar computational cost as training one model, we learn lines, curves, and simplexes of high-accuracy neural networks.
With a similar computational cost as training one model, we learn lines, curves, and simplexes of high-accuracy neural networks.
arXiv Detail & Related papers (2021-02-20T23:26:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.