Deep Transfer-Learning for patient specific model re-calibration:
Application to sEMG-Classification
- URL: http://arxiv.org/abs/2112.15019v1
- Date: Thu, 30 Dec 2021 11:35:53 GMT
- Title: Deep Transfer-Learning for patient specific model re-calibration:
Application to sEMG-Classification
- Authors: Stephan Johann Lehmler, Muhammad Saif-ur-Rehman, Tobias Glasmachers,
Ioannis Iossifidis
- Abstract summary: Machine learning based sEMG decoders are either trained on subject-specific data, or at least recalibrated for each user, individually.
Due to the limited amount of availability of sEMG data, the deep learning models are prone to overfitting.
Recently, transfer learning for domain adaptation improved generalization quality with reduced training time.
- Score: 0.2676349883103404
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accurate decoding of surface electromyography (sEMG) is pivotal for
muscle-to-machine-interfaces (MMI) and their application for e.g.
rehabilitation therapy. sEMG signals have high inter-subject variability, due
to various factors, including skin thickness, body fat percentage, and
electrode placement. Therefore, obtaining high generalization quality of a
trained sEMG decoder is quite challenging. Usually, machine learning based sEMG
decoders are either trained on subject-specific data, or at least recalibrated
for each user, individually. Even though, deep learning algorithms produced
several state of the art results for sEMG decoding,however, due to the limited
amount of availability of sEMG data, the deep learning models are prone to
overfitting. Recently, transfer learning for domain adaptation improved
generalization quality with reduced training time on various machine learning
tasks. In this study, we investigate the effectiveness of transfer learning
using weight initialization for recalibration of two different pretrained deep
learning models on a new subjects data, and compare their performance to
subject-specific models. To the best of our knowledge, this is the first study
that thoroughly investigated weight-initialization based transfer learning for
sEMG classification and compared transfer learning to subject-specific
modeling. We tested our models on three publicly available databases under
various settings. On average over all settings, our transfer learning approach
improves 5~\%-points on the pretrained models without fine-tuning and
12~\%-points on the subject-specific models, while being trained on average
22~\% fewer epochs. Our results indicate that transfer learning enables faster
training on fewer samples than user-specific models, and improves the
performance of pretrained models as long as enough data is available.
Related papers
- Exploring Learngene via Stage-wise Weight Sharing for Initializing Variable-sized Models [40.21274215353816]
We introduce the Learngene framework, which learns one compact part termed as learngene from a large well-trained model.
We then expand these learngene layers containing stage information at their corresponding stage to initialize models of variable depths.
Experiments on ImageNet-1K demonstrate that SWS achieves consistent better performance compared to many models trained from scratch.
arXiv Detail & Related papers (2024-04-25T06:04:34Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models [115.501751261878]
Fine-tuning language models(LMs) on human-generated data remains a prevalent practice.
We investigate whether we can go beyond human data on tasks where we have access to scalar feedback.
We find that ReST$EM$ scales favorably with model size and significantly surpasses fine-tuning only on human data.
arXiv Detail & Related papers (2023-12-11T18:17:43Z) - Reducing Intraspecies and Interspecies Covariate Shift in Traumatic
Brain Injury EEG of Humans and Mice Using Transfer Euclidean Alignment [4.264615907591813]
High variability across subjects poses a significant challenge when it comes to deploying machine learning models for classification tasks in the real world.
In such instances, machine learning models that exhibit exceptional performance on a specific dataset may not necessarily demonstrate similar proficiency when applied to a distinct dataset for the same task.
We introduce Transfer Euclidean Alignment - a transfer learning technique to tackle the problem of the robustness of human biomedical data for training deep learning models.
arXiv Detail & Related papers (2023-10-03T19:48:02Z) - Towards Foundation Models for Scientific Machine Learning:
Characterizing Scaling and Transfer Behavior [32.74388989649232]
We study how pre-training could be used for scientific machine learning (SciML) applications.
We find that fine-tuning these models yields more performance gains as model size increases.
arXiv Detail & Related papers (2023-06-01T00:32:59Z) - Core-set Selection Using Metrics-based Explanations (CSUME) for
multiclass ECG [2.0520503083305073]
We show how a selection of good quality data improves deep learning model performance.
Our experimental results show a 9.67% and 8.69% precision and recall improvement with a significant training data volume reduction of 50%.
arXiv Detail & Related papers (2022-05-28T19:36:28Z) - Model-Agnostic Multitask Fine-tuning for Few-shot Vision-Language
Transfer Learning [59.38343286807997]
We propose Model-Agnostic Multitask Fine-tuning (MAMF) for vision-language models on unseen tasks.
Compared with model-agnostic meta-learning (MAML), MAMF discards the bi-level optimization and uses only first-order gradients.
We show that MAMF consistently outperforms the classical fine-tuning method for few-shot transfer learning on five benchmark datasets.
arXiv Detail & Related papers (2022-03-09T17:26:53Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Evaluating deep transfer learning for whole-brain cognitive decoding [11.898286908882561]
Transfer learning (TL) is well-suited to improve the performance of deep learning (DL) models in datasets with small numbers of samples.
Here, we evaluate TL for the application of DL models to the decoding of cognitive states from whole-brain functional Magnetic Resonance Imaging (fMRI) data.
arXiv Detail & Related papers (2021-11-01T15:44:49Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - Deep transfer learning for improving single-EEG arousal detection [63.52264764099532]
Two datasets do not contain exactly the same setup leading to degraded performance in single-EEG models.
We train a baseline model and replace the first two layers to prepare the architecture for single-channel electroencephalography data.
Using a fine-tuning strategy, our model yields similar performance to the baseline model and was significantly better than a comparable single-channel model.
arXiv Detail & Related papers (2020-04-10T16:51:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.