A Correlation-Ratio Transfer Learning and Variational Stein's Paradox
- URL: http://arxiv.org/abs/2206.06086v1
- Date: Fri, 10 Jun 2022 01:59:16 GMT
- Title: A Correlation-Ratio Transfer Learning and Variational Stein's Paradox
- Authors: Lu Lin and Weiyu Li
- Abstract summary: This paper introduces a new strategy, linear correlation-ratio, to build an accurate relationship between the models.
On the practical side, the new framework is applied to some application scenarios, especially the areas of data streams and medical studies.
- Score: 7.652701739127332
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A basic condition for efficient transfer learning is the similarity between a
target model and source models. In practice, however, the similarity condition
is difficult to meet or is even violated. Instead of the similarity condition,
a brand-new strategy, linear correlation-ratio, is introduced in this paper to
build an accurate relationship between the models. Such a correlation-ratio can
be easily estimated by historical data or a part of sample. Then, a
correlation-ratio transfer learning likelihood is established based on the
correlation-ratio combination. On the practical side, the new framework is
applied to some application scenarios, especially the areas of data streams and
medical studies. Methodologically, some techniques are suggested for
transferring the information from simple source models to a relatively complex
target model. Theoretically, some favorable properties, including the global
convergence rate, are achieved, even for the case where the source models are
not similar to the target model. All in all, it can be seen from the theories
and experimental results that the inference on the target model is
significantly improved by the information from similar or dissimilar source
models. In other words, a variational Stein's paradox is illustrated in the
context of transfer learning.
Related papers
- Accelerated Stochastic ExtraGradient: Mixing Hessian and Gradient Similarity to Reduce Communication in Distributed and Federated Learning [50.382793324572845]
Distributed computing involves communication between devices, which requires solving two key problems: efficiency and privacy.
In this paper, we analyze a new method that incorporates the ideas of using data similarity and clients sampling.
To address privacy concerns, we apply the technique of additional noise and analyze its impact on the convergence of the proposed method.
arXiv Detail & Related papers (2024-09-22T00:49:10Z) - Comparing Foundation Models using Data Kernels [13.099029073152257]
We present a methodology for directly comparing the embedding space geometry of foundation models.
Our methodology is grounded in random graph theory and enables valid hypothesis testing of embedding similarity.
We show how our framework can induce a manifold of models equipped with a distance function that correlates strongly with several downstream metrics.
arXiv Detail & Related papers (2023-05-09T02:01:07Z) - Less is More: Mitigate Spurious Correlations for Open-Domain Dialogue
Response Generation Models by Causal Discovery [52.95935278819512]
We conduct the first study on spurious correlations for open-domain response generation models based on a corpus CGDIALOG curated in our work.
Inspired by causal discovery algorithms, we propose a novel model-agnostic method for training and inference of response generation model.
arXiv Detail & Related papers (2023-03-02T06:33:48Z) - On the Strong Correlation Between Model Invariance and Generalization [54.812786542023325]
Generalization captures a model's ability to classify unseen data.
Invariance measures consistency of model predictions on transformations of the data.
From a dataset-centric view, we find a certain model's accuracy and invariance linearly correlated on different test sets.
arXiv Detail & Related papers (2022-07-14T17:08:25Z) - Bias-inducing geometries: an exactly solvable data model with fairness
implications [13.690313475721094]
We introduce an exactly solvable high-dimensional model of data imbalance.
We analytically unpack the typical properties of learning models trained in this synthetic framework.
We obtain exact predictions for the observables that are commonly employed for fairness assessment.
arXiv Detail & Related papers (2022-05-31T16:27:57Z) - Similarity learning for wells based on logging data [8.265576412171702]
We propose a novel framework to solve the geological profile similarity estimation based on a deep learning model.
Our similarity model takes well-logging data as input and provides the similarity of wells as output.
For model testing, we used two open datasets originating in New Zealand and Norway.
arXiv Detail & Related papers (2022-02-11T12:47:56Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Split Modeling for High-Dimensional Logistic Regression [0.2676349883103404]
A novel method is proposed to an ensemble logistic classification model briefly compiled.
Our method learns how to exploit the bias-off resulting in excellent prediction accuracy.
An open-source software library implementing the proposed method is discussed.
arXiv Detail & Related papers (2021-02-17T05:57:26Z) - Few-shot Visual Reasoning with Meta-analogical Contrastive Learning [141.2562447971]
We propose to solve a few-shot (or low-shot) visual reasoning problem, by resorting to analogical reasoning.
We extract structural relationships between elements in both domains, and enforce them to be as similar as possible with analogical learning.
We validate our method on RAVEN dataset, on which it outperforms state-of-the-art method, with larger gains when the training data is scarce.
arXiv Detail & Related papers (2020-07-23T14:00:34Z) - Pairwise Supervision Can Provably Elicit a Decision Boundary [84.58020117487898]
Similarity learning is a problem to elicit useful representations by predicting the relationship between a pair of patterns.
We show that similarity learning is capable of solving binary classification by directly eliciting a decision boundary.
arXiv Detail & Related papers (2020-06-11T05:35:16Z) - Amortized Bayesian model comparison with evidential deep learning [0.12314765641075436]
We propose a novel method for performing Bayesian model comparison using specialized deep learning architectures.
Our method is purely simulation-based and circumvents the step of explicitly fitting all alternative models under consideration to each observed dataset.
We show that our method achieves excellent results in terms of accuracy, calibration, and efficiency across the examples considered in this work.
arXiv Detail & Related papers (2020-04-22T15:15:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.