Feasibility of Transfer Learning: A Mathematical Framework
- URL: http://arxiv.org/abs/2305.12985v1
- Date: Mon, 22 May 2023 12:44:38 GMT
- Title: Feasibility of Transfer Learning: A Mathematical Framework
- Authors: Haoyang Cao and Haotian Gu and Xin Guo
- Abstract summary: It begins by establishing the necessary mathematical concepts and constructing a mathematical framework for transfer learning.
It then identifies and formulates the three-step transfer learning procedure as an optimization problem, allowing for the resolution of the feasibility issue.
- Score: 4.530876736231948
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Transfer learning is a popular paradigm for utilizing existing knowledge from
previous learning tasks to improve the performance of new ones. It has enjoyed
numerous empirical successes and inspired a growing number of theoretical
studies. This paper addresses the feasibility issue of transfer learning. It
begins by establishing the necessary mathematical concepts and constructing a
mathematical framework for transfer learning. It then identifies and formulates
the three-step transfer learning procedure as an optimization problem, allowing
for the resolution of the feasibility issue. Importantly, it demonstrates that
under certain technical conditions, such as appropriate choice of loss
functions and data sets, an optimal procedure for transfer learning exists.
This study of the feasibility issue brings additional insights into various
transfer learning problems. It sheds light on the impact of feature
augmentation on model performance, explores potential extensions of domain
adaptation, and examines the feasibility of efficient feature extractor
transfer in image classification.
Related papers
- Risk of Transfer Learning and its Applications in Finance [2.966069495345018]
We propose a novel concept of transfer risk and analyze its properties to evaluate transferability of transfer learning.
Numerical results demonstrate a strong correlation between transfer risk and overall transfer learning performance.
arXiv Detail & Related papers (2023-11-06T17:23:54Z) - Feasibility and Transferability of Transfer Learning: A Mathematical
Framework [4.031388559887924]
We build for the first time a mathematical framework for the general procedure of transfer learning.
We also propose a novel concept of transfer risk to evaluate transferability of transfer learning.
arXiv Detail & Related papers (2023-01-27T05:54:53Z) - Learning to Transfer: A Foliated Theory [18.58482811176484]
Learning to transfer considers learning solutions to tasks in a such a way that relevant knowledge can be transferred to new, related tasks.
This is important for general learning, as well as for improving the efficiency of the learning process.
We introduce a framework using the differential geometric theory of foliations that provides such a foundation.
arXiv Detail & Related papers (2021-07-22T15:46:45Z) - Unsupervised Transfer Learning for Spatiotemporal Predictive Networks [90.67309545798224]
We study how to transfer knowledge from a zoo of unsupervisedly learned models towards another network.
Our motivation is that models are expected to understand complex dynamics from different sources.
Our approach yields significant improvements on three benchmarks fortemporal prediction, and benefits the target even from less relevant ones.
arXiv Detail & Related papers (2020-09-24T15:40:55Z) - Transfer Learning in Deep Reinforcement Learning: A Survey [64.36174156782333]
Reinforcement learning is a learning paradigm for solving sequential decision-making problems.
Recent years have witnessed remarkable progress in reinforcement learning upon the fast development of deep neural networks.
transfer learning has arisen to tackle various challenges faced by reinforcement learning.
arXiv Detail & Related papers (2020-09-16T18:38:54Z) - Importance Weighted Policy Learning and Adaptation [89.46467771037054]
We study a complementary approach which is conceptually simple, general, modular and built on top of recent improvements in off-policy learning.
The framework is inspired by ideas from the probabilistic inference literature and combines robust off-policy learning with a behavior prior.
Our approach achieves competitive adaptation performance on hold-out tasks compared to meta reinforcement learning baselines and can scale to complex sparse-reward scenarios.
arXiv Detail & Related papers (2020-09-10T14:16:58Z) - What is being transferred in transfer learning? [51.6991244438545]
We show that when training from pre-trained weights, the model stays in the same basin in the loss landscape.
We present that when training from pre-trained weights, the model stays in the same basin in the loss landscape and different instances of such model are similar in feature space and close in parameter space.
arXiv Detail & Related papers (2020-08-26T17:23:40Z) - Uniform Priors for Data-Efficient Transfer [65.086680950871]
We show that features that are most transferable have high uniformity in the embedding space.
We evaluate the regularization on its ability to facilitate adaptation to unseen tasks and data.
arXiv Detail & Related papers (2020-06-30T04:39:36Z) - Limits of Transfer Learning [0.0]
We show the need to carefully select which sets of information to transfer and the need for dependence between transferred information and target problems.
These results build on the algorithmic search framework for machine learning, allowing the results to apply to a wide range of learning problems using transfer.
arXiv Detail & Related papers (2020-06-23T01:48:23Z) - Task-Feature Collaborative Learning with Application to Personalized
Attribute Prediction [166.87111665908333]
We propose a novel multi-task learning method called Task-Feature Collaborative Learning (TFCL)
Specifically, we first propose a base model with a heterogeneous block-diagonal structure regularizer to leverage the collaborative grouping of features and tasks.
As a practical extension, we extend the base model by allowing overlapping features and differentiating the hard tasks.
arXiv Detail & Related papers (2020-04-29T02:32:04Z) - Inter- and Intra-domain Knowledge Transfer for Related Tasks in Deep
Character Recognition [2.320417845168326]
Pre-training a deep neural network on the ImageNet dataset is a common practice for training deep learning models.
The technique of pre-training on one task and then retraining on a new one is called transfer learning.
In this paper we analyse the effectiveness of using deep transfer learning for character recognition tasks.
arXiv Detail & Related papers (2020-01-02T14:18:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.