Exploring the Efficacy of Transfer Learning in Mining Image-Based
Software Artifacts
- URL: http://arxiv.org/abs/2003.01627v1
- Date: Tue, 3 Mar 2020 16:41:45 GMT
- Title: Exploring the Efficacy of Transfer Learning in Mining Image-Based
Software Artifacts
- Authors: Natalie Best, Jordan Ott, Erik Linstead
- Abstract summary: Transfer learning allows us to train deep architectures requiring a large number of learned parameters, even if the amount of available data is limited.
Here we explore the applicability of transfer learning utilizing models pre-trained on non-software engineering data applied to the problem of classifying software diagrams.
- Score: 1.5285292154680243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transfer learning allows us to train deep architectures requiring a large
number of learned parameters, even if the amount of available data is limited,
by leveraging existing models previously trained for another task. Here we
explore the applicability of transfer learning utilizing models pre-trained on
non-software engineering data applied to the problem of classifying software
UML diagrams. Our experimental results show training reacts positively to
transfer learning as related to sample size, even though the pre-trained model
was not exposed to training instances from the software domain. We contrast the
transferred network with other networks to show its advantage on different
sized training sets, which indicates that transfer learning is equally
effective to custom deep architectures when large amounts of training data is
not available.
Related papers
- Encapsulating Knowledge in One Prompt [56.31088116526825]
KiOP encapsulates knowledge from various models into a solitary prompt without altering the original models or requiring access to the training data.
From a practicality standpoint, this paradigm proves the effectiveness of Visual Prompt in data inaccessible contexts.
Experiments across various datasets and models demonstrate the efficacy of the proposed KiOP knowledge transfer paradigm.
arXiv Detail & Related papers (2024-07-16T16:35:23Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - Novel transfer learning schemes based on Siamese networks and synthetic
data [6.883906273999368]
Transfer learning schemes based on deep networks offer state-of-the-art technologies in computer vision.
Such applications are currently restricted to application domains where suitable deepnetwork models are readily available.
We propose a novel transfer learning scheme which expands a recently introduced Twin-VAE architecture.
arXiv Detail & Related papers (2022-11-21T09:48:21Z) - Training Deep Networks from Zero to Hero: avoiding pitfalls and going
beyond [59.94347858883343]
This tutorial covers the basic steps as well as more recent options to improve models.
It can be particularly useful in datasets that are not as well-prepared as those in challenges.
arXiv Detail & Related papers (2021-09-06T21:31:42Z) - Transfer of Pretrained Model Weights Substantially Improves
Semi-Supervised Image Classification [3.492636597449942]
Deep neural networks produce state-of-the-art results when trained on a large number of labeled examples.
Deep neural networks tend to overfit when small amounts of labeled examples are used for training.
We show that transfer learning always substantially improves the model's accuracy when few labeled examples are available.
arXiv Detail & Related papers (2021-09-02T08:58:34Z) - What is being transferred in transfer learning? [51.6991244438545]
We show that when training from pre-trained weights, the model stays in the same basin in the loss landscape.
We present that when training from pre-trained weights, the model stays in the same basin in the loss landscape and different instances of such model are similar in feature space and close in parameter space.
arXiv Detail & Related papers (2020-08-26T17:23:40Z) - Adversarially-Trained Deep Nets Transfer Better: Illustration on Image
Classification [53.735029033681435]
Transfer learning is a powerful methodology for adapting pre-trained deep neural networks on image recognition tasks to new domains.
In this work, we demonstrate that adversarially-trained models transfer better than non-adversarially-trained models.
arXiv Detail & Related papers (2020-07-11T22:48:42Z) - Minimax Lower Bounds for Transfer Learning with Linear and One-hidden
Layer Neural Networks [27.44348371795822]
We develop a statistical minimax framework to characterize the limits of transfer learning.
We derive a lower-bound for the target generalization error achievable by any algorithm as a function of the number of labeled source and target data.
arXiv Detail & Related papers (2020-06-16T22:49:26Z) - The Utility of Feature Reuse: Transfer Learning in Data-Starved Regimes [6.419457653976053]
We describe a transfer learning use case for a domain with a data-starved regime.
We evaluate the effectiveness of convolutional feature extraction and fine-tuning.
We conclude that transfer learning enhances the performance of CNN architectures in data-starved regimes.
arXiv Detail & Related papers (2020-02-29T18:48:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.