Novel transfer learning schemes based on Siamese networks and synthetic
data
- URL: http://arxiv.org/abs/2211.11308v2
- Date: Tue, 22 Nov 2022 13:14:41 GMT
- Title: Novel transfer learning schemes based on Siamese networks and synthetic
data
- Authors: Dominik Stallmann and Philip Kenneweg and Barbara Hammer
- Abstract summary: Transfer learning schemes based on deep networks offer state-of-the-art technologies in computer vision.
Such applications are currently restricted to application domains where suitable deepnetwork models are readily available.
We propose a novel transfer learning scheme which expands a recently introduced Twin-VAE architecture.
- Score: 6.883906273999368
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Transfer learning schemes based on deep networks which have been trained on
huge image corpora offer state-of-the-art technologies in computer vision.
Here, supervised and semi-supervised approaches constitute efficient
technologies which work well with comparably small data sets. Yet, such
applications are currently restricted to application domains where suitable
deepnetwork models are readily available. In this contribution, we address an
important application area in the domain of biotechnology, the automatic
analysis of CHO-K1 suspension growth in microfluidic single-cell cultivation,
where data characteristics are very dissimilar to existing domains and trained
deep networks cannot easily be adapted by classical transfer learning. We
propose a novel transfer learning scheme which expands a recently introduced
Twin-VAE architecture, which is trained on realistic and synthetic data, and we
modify its specialized training procedure to the transfer learning domain. In
the specific domain, often only few to no labels exist and annotations are
costly. We investigate a novel transfer learning strategy, which incorporates a
simultaneous retraining on natural and synthetic data using an invariant shared
representation as well as suitable target variables, while it learns to handle
unseen data from a different microscopy tech nology. We show the superiority of
the variation of our Twin-VAE architecture over the state-of-the-art transfer
learning methodology in image processing as well as classical image processing
technologies, which persists, even with strongly shortened training times and
leads to satisfactory results in this domain. The source code is available at
https://github.com/dstallmann/transfer_learning_twinvae, works cross-platform,
is open-source and free (MIT licensed) software. We make the data sets
available at https://pub.uni-bielefeld.de/record/2960030.
Related papers
- Unsupervised Domain Transfer with Conditional Invertible Neural Networks [83.90291882730925]
We propose a domain transfer approach based on conditional invertible neural networks (cINNs)
Our method inherently guarantees cycle consistency through its invertible architecture, and network training can efficiently be conducted with maximum likelihood.
Our method enables the generation of realistic spectral data and outperforms the state of the art on two downstream classification tasks.
arXiv Detail & Related papers (2023-03-17T18:00:27Z) - ProCST: Boosting Semantic Segmentation using Progressive Cyclic
Style-Transfer [38.03127458140549]
We propose a novel two-stage framework for improving domain adaptation techniques.
In the first step, we progressively train a multi-scale neural network to perform an initial transfer between the source data to the target data.
This new data has a reduced domain gap from the desired target domain, and the applied UDA approach further closes the gap.
arXiv Detail & Related papers (2022-04-25T18:01:05Z) - Self-Supervised Graph Neural Network for Multi-Source Domain Adaptation [51.21190751266442]
Domain adaptation (DA) tries to tackle the scenarios when the test data does not fully follow the same distribution of the training data.
By learning from large-scale unlabeled samples, self-supervised learning has now become a new trend in deep learning.
We propose a novel textbfSelf-textbfSupervised textbfGraph Neural Network (SSG) to enable more effective inter-task information exchange and knowledge sharing.
arXiv Detail & Related papers (2022-04-08T03:37:56Z) - DAFormer: Improving Network Architectures and Training Strategies for
Domain-Adaptive Semantic Segmentation [99.88539409432916]
We study the unsupervised domain adaptation (UDA) process.
We propose a novel UDA method, DAFormer, based on the benchmark results.
DAFormer significantly improves the state-of-the-art performance by 10.8 mIoU for GTA->Cityscapes and 5.4 mIoU for Synthia->Cityscapes.
arXiv Detail & Related papers (2021-11-29T19:00:46Z) - Unsupervised Cross-Domain Prerequisite Chain Learning using Variational
Graph Autoencoders [2.735701323590668]
We propose unsupervised cross-domain concept prerequisite chain learning using an optimized variational graph autoencoder.
Our model learns to transfer concept prerequisite relations from an information-rich domain to an information-poor domain.
Also, we expand an existing dataset by introducing two new domains: CV and Bioinformatics.
arXiv Detail & Related papers (2021-05-07T21:02:41Z) - Flexible deep transfer learning by separate feature embeddings and
manifold alignment [0.0]
Object recognition is a key enabler across industry and defense.
Unfortunately, algorithms trained on existing labeled datasets do not directly generalize to new data because the data distributions do not match.
We propose a novel deep learning framework that overcomes this limitation by learning separate feature extractions for each domain.
arXiv Detail & Related papers (2020-12-22T19:24:44Z) - CosSGD: Nonlinear Quantization for Communication-efficient Federated
Learning [62.65937719264881]
Federated learning facilitates learning across clients without transferring local data on these clients to a central server.
We propose a nonlinear quantization for compressed gradient descent, which can be easily utilized in federated learning.
Our system significantly reduces the communication cost by up to three orders of magnitude, while maintaining convergence and accuracy of the training process.
arXiv Detail & Related papers (2020-12-15T12:20:28Z) - What is being transferred in transfer learning? [51.6991244438545]
We show that when training from pre-trained weights, the model stays in the same basin in the loss landscape.
We present that when training from pre-trained weights, the model stays in the same basin in the loss landscape and different instances of such model are similar in feature space and close in parameter space.
arXiv Detail & Related papers (2020-08-26T17:23:40Z) - CrossTransformers: spatially-aware few-shot transfer [92.33252608837947]
Given new tasks with very little data, modern vision systems degrade remarkably quickly.
We show how the neural network representations which underpin modern vision systems are subject to supervision collapse.
We propose self-supervised learning to encourage general-purpose features that transfer better.
arXiv Detail & Related papers (2020-07-22T15:37:08Z) - Exploring the Efficacy of Transfer Learning in Mining Image-Based
Software Artifacts [1.5285292154680243]
Transfer learning allows us to train deep architectures requiring a large number of learned parameters, even if the amount of available data is limited.
Here we explore the applicability of transfer learning utilizing models pre-trained on non-software engineering data applied to the problem of classifying software diagrams.
arXiv Detail & Related papers (2020-03-03T16:41:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.