Manually Selecting The Data Function for Supervised Learning of small
datasets
- URL: http://arxiv.org/abs/2303.03894v1
- Date: Tue, 7 Mar 2023 13:38:04 GMT
- Title: Manually Selecting The Data Function for Supervised Learning of small
datasets
- Authors: Amir Khanjari and Saeid Pourmand and Mohammad Reza Faridrohani
- Abstract summary: Supervised learning problems may become ill-posed when there is a lack of information.
Initializing an informative ill-posed operator is akin to posing better questions to achieve more accurate answers.
The Fredholm integral equation of the first kind (FIFK) is a reliable ill-posed operator that can integrate distributions and prior knowledge as input information.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Supervised learning problems may become ill-posed when there is a lack of
information, resulting in unstable and non-unique solutions. However, instead
of solely relying on regularization, initializing an informative ill-posed
operator is akin to posing better questions to achieve more accurate answers.
The Fredholm integral equation of the first kind (FIFK) is a reliable ill-posed
operator that can integrate distributions and prior knowledge as input
information. By incorporating input distributions and prior knowledge, the FIFK
operator can address the limitations of using high-dimensional input
distributions by semi-supervised assumptions, leading to more precise
approximations of the integral operator. Additionally, the FIFK's incorporation
of probabilistic principles can further enhance the accuracy and effectiveness
of solutions. In cases of noisy operator equations and limited data, the FIFK's
flexibility in defining problems using prior information or cross-validation
with various kernel designs is especially advantageous. This capability allows
for detailed problem definitions and facilitates achieving high levels of
accuracy and stability in solutions. In our study, we examined the FIFK through
two different approaches. Firstly, we implemented a semi-supervised assumption
by using the same Fredholm operator kernel and data function kernel and
incorporating unlabeled information. Secondly, we used the MSDF method, which
involves selecting different kernels on both sides of the equation to define
when the mapping kernel is different from the data function kernel. To assess
the effectiveness of the FIFK and the proposed methods in solving ill-posed
problems, we conducted experiments on a real-world dataset. Our goal was to
compare the performance of these methods against the widely used least-squares
method and other comparable methods.
Related papers
- Joint Optimization of Resource Allocation and Data Selection for Fast and Cost-Efficient Federated Edge Learning [9.460012379815423]
Deploying learning at the wireless edge introduces FEEL (FEEL)
We propose an efficient system, a federated joint resource allocation and data selection.
The superiority of our proposed scheme of joint resource allocation and data selection is validated.
arXiv Detail & Related papers (2024-07-03T08:03:59Z) - Efficient Prior Calibration From Indirect Data [5.588334720483076]
This paper is concerned with learning the prior model from data, in particular, learning the prior from multiple realizations of indirect data obtained through the noisy observation process.
An efficient residual-based neural operator approximation of the forward model is proposed and it is shown that this may be learned concurrently with the pushforward map.
arXiv Detail & Related papers (2024-05-28T08:34:41Z) - Data Imputation by Pursuing Better Classification: A Supervised Kernel-Based Method [23.16359277296206]
We propose a new framework that effectively leverages supervision information to complete missing data in a manner conducive to classification.
Our algorithm significantly outperforms other methods when the data is missing more than 60% of the features.
arXiv Detail & Related papers (2024-05-13T14:44:02Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Local Learning Matters: Rethinking Data Heterogeneity in Federated
Learning [61.488646649045215]
Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices)
arXiv Detail & Related papers (2021-11-28T19:03:39Z) - Numerical Approximation in CFD Problems Using Physics Informed Machine
Learning [0.0]
The thesis focuses on various techniques to find an alternate approximation method that could be universally used for a wide range of CFD problems.
The focus stays over physics informed machine learning techniques where solving differential equations is possible without any training with computed data.
The extreme learning machine (ELM) is a very fast neural network algorithm at the cost of tunable parameters.
arXiv Detail & Related papers (2021-11-01T22:54:51Z) - Offline Model-Based Optimization via Normalized Maximum Likelihood
Estimation [101.22379613810881]
We consider data-driven optimization problems where one must maximize a function given only queries at a fixed set of points.
This problem setting emerges in many domains where function evaluation is a complex and expensive process.
We propose a tractable approximation that allows us to scale our method to high-capacity neural network models.
arXiv Detail & Related papers (2021-02-16T06:04:27Z) - Federated Doubly Stochastic Kernel Learning for Vertically Partitioned
Data [93.76907759950608]
We propose a doubly kernel learning algorithm for vertically partitioned data.
We show that FDSKL is significantly faster than state-of-the-art federated learning methods when dealing with kernels.
arXiv Detail & Related papers (2020-08-14T05:46:56Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - Focus of Attention Improves Information Transfer in Visual Features [80.22965663534556]
This paper focuses on unsupervised learning for transferring visual information in a truly online setting.
The computation of the entropy terms is carried out by a temporal process which yields online estimation of the entropy terms.
In order to better structure the input probability distribution, we use a human-like focus of attention model.
arXiv Detail & Related papers (2020-06-16T15:07:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.