Transferring model structure in Bayesian transfer learning for Gaussian
process regression
- URL: http://arxiv.org/abs/2101.06884v1
- Date: Mon, 18 Jan 2021 05:28:02 GMT
- Title: Transferring model structure in Bayesian transfer learning for Gaussian
process regression
- Authors: Milan Pape\v{z}, Anthony Quinn
- Abstract summary: This paper defines the task of conditioning a target probability distribution on a transferred source distribution.
Fully probabilistic design is adopted to solve this optimal decision-making problem in the target.
By successfully transferring higher moments of the source, the target can reject unreliable source knowledge.
- Score: 1.370633147306388
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bayesian transfer learning (BTL) is defined in this paper as the task of
conditioning a target probability distribution on a transferred source
distribution. The target globally models the interaction between the source and
target, and conditions on a probabilistic data predictor made available by an
independent local source modeller. Fully probabilistic design is adopted to
solve this optimal decision-making problem in the target. By successfully
transferring higher moments of the source, the target can reject unreliable
source knowledge (i.e. it achieves robust transfer). This dual-modeller
framework means that the source's local processing of raw data into a
transferred predictive distribution -- with compressive possibilities -- is
enriched by (the possible expertise of) the local source model. In addition,
the introduction of the global target modeller allows correlation between the
source and target tasks -- if known to the target -- to be accounted for.
Important consequences emerge. Firstly, the new scheme attains the performance
of fully modelled (i.e. conventional) multitask learning schemes in (those
rare) cases where target model misspecification is avoided. Secondly, and more
importantly, the new dual-modeller framework is robust to the model
misspecification that undermines conventional multitask learning. We thoroughly
explore these issues in the key context of interacting Gaussian process
regression tasks. Experimental evidence from both synthetic and real data
settings validates our technical findings: that the proposed BTL framework
enjoys robustness in transfer while also being robust to model
misspecification.
Related papers
- Theoretically Guaranteed Distribution Adaptable Learning [23.121014921407898]
We propose a novel framework called Distribution Adaptable Learning (DAL)
DAL enables the model to effectively track the evolving data distributions.
It can enhance the reusable and evolvable properties of DAL in accommodating evolving distributions.
arXiv Detail & Related papers (2024-11-05T09:10:39Z) - Enhancing Accuracy in Generative Models via Knowledge Transfer [4.3108820946281945]
We introduce a novel framework for transfer learning under distribution metrics such as the Kullback-Leibler divergence.
Our theory suggests that the shared structures can augment the generation accuracy for a target task.
arXiv Detail & Related papers (2024-05-27T05:10:49Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Building a Winning Team: Selecting Source Model Ensembles using a
Submodular Transferability Estimation Approach [20.86345962679122]
Estimating the transferability of publicly available pretrained models to a target task has assumed an important place for transfer learning tasks.
We propose a novel Optimal tranSport-based suBmOdular tRaNsferability metric (OSBORN) to estimate the transferability of an ensemble of models to a downstream task.
arXiv Detail & Related papers (2023-09-05T17:57:31Z) - Towards Estimating Transferability using Hard Subsets [25.86053764521497]
We propose HASTE, a new strategy to estimate the transferability of a source model to a particular target task using only a harder subset of target data.
We show that HASTE can be used with any existing transferability metric to improve their reliability.
Our experimental results across multiple source model architectures, target datasets, and transfer learning tasks show that HASTE modified metrics are consistently better or on par with the state of the art transferability metrics.
arXiv Detail & Related papers (2023-01-17T14:50:18Z) - Variational Model Perturbation for Source-Free Domain Adaptation [64.98560348412518]
We introduce perturbations into the model parameters by variational Bayesian inference in a probabilistic framework.
We demonstrate the theoretical connection to learning Bayesian neural networks, which proves the generalizability of the perturbed model to target domains.
arXiv Detail & Related papers (2022-10-19T08:41:19Z) - Uncertainty-guided Source-free Domain Adaptation [77.3844160723014]
Source-free domain adaptation (SFDA) aims to adapt a classifier to an unlabelled target data set by only using a pre-trained source model.
We propose quantifying the uncertainty in the source model predictions and utilizing it to guide the target adaptation.
arXiv Detail & Related papers (2022-08-16T08:03:30Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Fully probabilistic design for knowledge fusion between Bayesian filters
under uniform disturbances [0.0]
This paper considers the problem of Bayesian transfer learning-based knowledge fusion between linear state-space processes.
A joint model of the target and source(s) is not required and is not elicited.
The resulting FPD-optimal target learner is robust, in the sense that it can reject poor-quality source knowledge.
arXiv Detail & Related papers (2021-09-22T08:49:15Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Universal Source-Free Domain Adaptation [57.37520645827318]
We propose a novel two-stage learning process for domain adaptation.
In the Procurement stage, we aim to equip the model for future source-free deployment, assuming no prior knowledge of the upcoming category-gap and domain-shift.
In the Deployment stage, the goal is to design a unified adaptation algorithm capable of operating across a wide range of category-gaps.
arXiv Detail & Related papers (2020-04-09T07:26:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.