High-resolution signal recovery via generalized sampling and functional
principal component analysis
- URL: http://arxiv.org/abs/2002.08724v3
- Date: Thu, 14 Oct 2021 16:54:33 GMT
- Title: High-resolution signal recovery via generalized sampling and functional
principal component analysis
- Authors: Milana Gataric
- Abstract summary: We introduce a computational framework for recovering a high-resolution approximation of an unknown function from its indirect measurements.
In particular, we increase the signal resolution via a data driven approach, which models the function of interest as a realization of a random field.
We show that the size of the training set can be reduced by leveraging sparse representations of the functional principal components.
- Score: 0.609170287691728
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we introduce a computational framework for recovering a
high-resolution approximation of an unknown function from its low-resolution
indirect measurements as well as high-resolution training observations by
merging the frameworks of generalized sampling and functional principal
component analysis. In particular, we increase the signal resolution via a data
driven approach, which models the function of interest as a realization of a
random field and leverages a training set of observations generated via the
same underlying random process. We study the performance of the resulting
estimation procedure and show that high-resolution recovery is indeed possible
provided appropriate low-rank and angle conditions hold and provided the
training set is sufficiently large relative to the desired resolution.
Moreover, we show that the size of the training set can be reduced by
leveraging sparse representations of the functional principal components.
Furthermore, the effectiveness of the proposed reconstruction procedure is
illustrated by various numerical examples.
Related papers
- On Discriminative Probabilistic Modeling for Self-Supervised Representation Learning [85.75164588939185]
We study the discriminative probabilistic modeling problem on a continuous domain for (multimodal) self-supervised representation learning.
We conduct generalization error analysis to reveal the limitation of current InfoNCE-based contrastive loss for self-supervised representation learning.
arXiv Detail & Related papers (2024-10-11T18:02:46Z) - FNP: Fourier Neural Processes for Arbitrary-Resolution Data Assimilation [58.149902193341816]
We propose textittextbfFourier Neural Processes (FNP) for textitarbitrary-resolution data assimilation in this paper.
Our FNP trained on a fixed resolution can directly handle the assimilation of observations with out-of-distribution resolutions and the observational information reconstruction task without additional fine-tuning.
arXiv Detail & Related papers (2024-06-03T12:24:24Z) - Stochastic Gradient Descent for Nonparametric Regression [11.24895028006405]
This paper introduces an iterative algorithm for training nonparametric additive models.
We show that the resulting inequality satisfies an oracle that allows for model mis-specification.
arXiv Detail & Related papers (2024-01-01T08:03:52Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - A General Framework for Sample-Efficient Function Approximation in
Reinforcement Learning [132.45959478064736]
We propose a general framework that unifies model-based and model-free reinforcement learning.
We propose a novel estimation function with decomposable structural properties for optimization-based exploration.
Under our framework, a new sample-efficient algorithm namely OPtimization-based ExploRation with Approximation (OPERA) is proposed.
arXiv Detail & Related papers (2022-09-30T17:59:16Z) - Reinforcement Learning from Partial Observation: Linear Function Approximation with Provable Sample Efficiency [111.83670279016599]
We study reinforcement learning for partially observed decision processes (POMDPs) with infinite observation and state spaces.
We make the first attempt at partial observability and function approximation for a class of POMDPs with a linear structure.
arXiv Detail & Related papers (2022-04-20T21:15:38Z) - A Forward Backward Greedy approach for Sparse Multiscale Learning [0.0]
We propose a feature driven Reproducing Kernel Hilbert space (RKHS) for which the associated kernel has a weighted multiscale structure.
For generating approximations in this space, we provide a practical forward-backward algorithm that is shown to greedily construct a set of basis functions having a multiscale structure.
We analyze the performance of the approach on a variety of simulation and real data sets.
arXiv Detail & Related papers (2021-02-14T04:22:52Z) - Feature space approximation for kernel-based supervised learning [2.653409741248232]
The goal is to reduce the size of the training data, resulting in lower storage consumption and computational complexity.
We demonstrate significant improvements in comparison to the computation of data-driven predictions involving the full training data set.
The method is applied to classification and regression problems from different application areas such as image recognition, system identification, and oceanographic time series analysis.
arXiv Detail & Related papers (2020-11-25T11:23:58Z) - Loss Bounds for Approximate Influence-Based Abstraction [81.13024471616417]
Influence-based abstraction aims to gain leverage by modeling local subproblems together with the 'influence' that the rest of the system exerts on them.
This paper investigates the performance of such approaches from a theoretical perspective.
We show that neural networks trained with cross entropy are well suited to learn approximate influence representations.
arXiv Detail & Related papers (2020-11-03T15:33:10Z) - A Support Detection and Root Finding Approach for Learning
High-dimensional Generalized Linear Models [10.103666349083165]
We develop a support detection and root finding procedure to learn the high dimensional sparse generalized linear models.
We conduct simulations and real data analysis to illustrate the advantages of our proposed method over several existing methods.
arXiv Detail & Related papers (2020-01-16T14:35:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.