Empirically Measuring Transfer Distance for System Design and Operation
- URL: http://arxiv.org/abs/2107.01184v1
- Date: Fri, 2 Jul 2021 16:45:58 GMT
- Title: Empirically Measuring Transfer Distance for System Design and Operation
- Authors: Tyler Cody, Stephen Adams, Peter A. Beling
- Abstract summary: We show that transfer learning algorithms have little, if any, examples from which to learn.
We consider the use of transfer distance in the design of machine rebuild procedures to allow for transferable prognostic models.
Practitioners can use the presented methodology to design and operate systems with consideration for the learning theoretic challenges faced by component learning systems.
- Score: 2.9864637081333085
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classical machine learning approaches are sensitive to non-stationarity.
Transfer learning can address non-stationarity by sharing knowledge from one
system to another, however, in areas like machine prognostics and defense, data
is fundamentally limited. Therefore, transfer learning algorithms have little,
if any, examples from which to learn. Herein, we suggest that these constraints
on algorithmic learning can be addressed by systems engineering. We formally
define transfer distance in general terms and demonstrate its use in
empirically quantifying the transferability of models. We consider the use of
transfer distance in the design of machine rebuild procedures to allow for
transferable prognostic models. We also consider the use of transfer distance
in predicting operational performance in computer vision. Practitioners can use
the presented methodology to design and operate systems with consideration for
the learning theoretic challenges faced by component learning systems.
Related papers
- Transfer Learning Study of Motion Transformer-based Trajectory Predictions [1.2972104025246092]
Trajectory planning in autonomous driving is highly dependent on predicting the emergent behavior of other road users.
Learning-based methods are currently showing impressive results in simulation-based challenges.
The study aims to provide insights into possible trade-offs between computational time and performance to support effective transfers into the real world.
arXiv Detail & Related papers (2024-04-12T06:50:32Z) - Interpretable Meta-Learning of Physical Systems [4.343110120255532]
Recent meta-learning methods rely on black-box neural networks, resulting in high computational costs and limited interpretability.
We argue that multi-environment generalization can be achieved using a simpler learning model, with an affine structure with respect to the learning task.
We demonstrate the competitive generalization performance and the low computational cost of our method by comparing it to state-of-the-art algorithms on physical systems.
arXiv Detail & Related papers (2023-12-01T10:18:50Z) - Learning Physical Concepts in Cyber-Physical Systems: A Case Study [72.74318982275052]
We provide an overview of the current state of research regarding methods for learning physical concepts in time series data.
We also analyze the most important methods from the current state of the art using the example of a three-tank system.
arXiv Detail & Related papers (2021-11-28T14:24:52Z) - Learning-Based UE Classification in Millimeter-Wave Cellular Systems
With Mobility [67.81523988596841]
Millimeter-wave cellular communication requires beamforming procedures that enable alignment of the transmitter and receiver beams as the user equipment (UE) moves.
For efficient beam tracking it is advantageous to classify users according to their traffic and mobility patterns.
Research to date has demonstrated efficient ways of machine learning based UE classification.
arXiv Detail & Related papers (2021-09-13T12:00:45Z) - A Bayesian Approach to (Online) Transfer Learning: Theory and Algorithms [6.193838300896449]
We study transfer learning from a Bayesian perspective, where a parametric statistical model is used.
Specifically, we study three variants of transfer learning problems, instantaneous, online, and time-variant transfer learning.
For each problem, we define an appropriate objective function, and provide either exact expressions or upper bounds on the learning performance.
Examples show that the derived bounds are accurate even for small sample sizes.
arXiv Detail & Related papers (2021-09-03T08:43:29Z) - An Extensible Benchmark Suite for Learning to Simulate Physical Systems [60.249111272844374]
We introduce a set of benchmark problems to take a step towards unified benchmarks and evaluation protocols.
We propose four representative physical systems, as well as a collection of both widely used classical time-based and representative data-driven methods.
arXiv Detail & Related papers (2021-08-09T17:39:09Z) - Federated Learning: A Signal Processing Perspective [144.63726413692876]
Federated learning is an emerging machine learning paradigm for training models across multiple edge devices holding local datasets, without explicitly exchanging the data.
This article provides a unified systematic framework for federated learning in a manner that encapsulates and highlights the main challenges that are natural to treat using signal processing tools.
arXiv Detail & Related papers (2021-03-31T15:14:39Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z) - Transfer Learning as an Enabler of the Intelligent Digital Twin [0.0]
Digital Twins have been described as beneficial in many areas, such as virtual commissioning, fault prediction or reconfiguration planning.
This article presents several cross-phase industrial transfer learning use cases utilizing intelligent Digital Twins.
arXiv Detail & Related papers (2020-12-03T13:51:05Z) - What is being transferred in transfer learning? [51.6991244438545]
We show that when training from pre-trained weights, the model stays in the same basin in the loss landscape.
We present that when training from pre-trained weights, the model stays in the same basin in the loss landscape and different instances of such model are similar in feature space and close in parameter space.
arXiv Detail & Related papers (2020-08-26T17:23:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.