Generalization of Fitness Exercise Recognition from Doppler Measurements
by Domain-adaption and Few-Shot Learning
- URL: http://arxiv.org/abs/2311.11910v1
- Date: Mon, 20 Nov 2023 16:40:48 GMT
- Title: Generalization of Fitness Exercise Recognition from Doppler Measurements
by Domain-adaption and Few-Shot Learning
- Authors: Biying Fu, Naser Damer, Florian Kirchbuchner, and Arjan Kuijper
- Abstract summary: In previous works, a mobile application was developed using an unmodified commercial off-the-shelf smartphone to recognize whole-body exercises.
Applying such a lab-environment trained model on realistic application variations causes a significant drop in performance.
This paper presents a database with controlled and uncontrolled subsets of fitness exercises.
- Score: 12.238586191793997
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In previous works, a mobile application was developed using an unmodified
commercial off-the-shelf smartphone to recognize whole-body exercises. The
working principle was based on the ultrasound Doppler sensing with the device
built-in hardware. Applying such a lab-environment trained model on realistic
application variations causes a significant drop in performance, and thus
decimate its applicability. The reason of the reduced performance can be
manifold. It could be induced by the user, environment, and device variations
in realistic scenarios. Such scenarios are often more complex and diverse,
which can be challenging to anticipate in the initial training data. To study
and overcome this issue, this paper presents a database with controlled and
uncontrolled subsets of fitness exercises. We propose two concepts to utilize
small adaption data to successfully improve model generalization in an
uncontrolled environment, increasing the recognition accuracy by two to six
folds compared to the baseline for different users.
Related papers
- A Practitioner's Guide to Continual Multimodal Pretraining [83.63894495064855]
Multimodal foundation models serve numerous applications at the intersection of vision and language.
To keep models updated, research into continual pretraining mainly explores scenarios with either infrequent, indiscriminate updates on large-scale new data, or frequent, sample-level updates.
We introduce FoMo-in-Flux, a continual multimodal pretraining benchmark with realistic compute constraints and practical deployment requirements.
arXiv Detail & Related papers (2024-08-26T17:59:01Z) - What Makes Pre-Trained Visual Representations Successful for Robust
Manipulation? [57.92924256181857]
We find that visual representations designed for manipulation and control tasks do not necessarily generalize under subtle changes in lighting and scene texture.
We find that emergent segmentation ability is a strong predictor of out-of-distribution generalization among ViT models.
arXiv Detail & Related papers (2023-11-03T18:09:08Z) - Modality-invariant Visual Odometry for Embodied Vision [1.7188280334580197]
Visual Odometry (VO) is a practical substitute for unreliable GPS and compass sensors.
Recent deep VO models limit themselves to a fixed set of input modalities, e.g., RGB and depth, while training on millions of samples.
We propose a Transformer-based modality-invariant VO approach that can deal with diverse or changing sensor suites of navigation agents.
arXiv Detail & Related papers (2023-04-29T21:47:12Z) - Versatile User Identification in Extended Reality using Pretrained Similarity-Learning [16.356961801884562]
We develop a similarity-learning model and pretrained it on the "Who Is Alyx?" dataset.
In comparison with a traditional classification-learning baseline, our model shows superior performance.
Our approach paves the way for easy integration of pretrained motion-based identification models in production XR systems.
arXiv Detail & Related papers (2023-02-15T08:26:24Z) - Reducing Impacts of System Heterogeneity in Federated Learning using
Weight Update Magnitudes [0.0]
Federated learning enables machine learning models to train locally on each handheld device while only synchronizing their neuron updates with a server.
This results in the training time of federated learning tasks being dictated by a few low-performance straggler devices.
In this work, we aim to mitigate the performance bottleneck of federated learning by dynamically forming sub-models for stragglers based on their performance and accuracy feedback.
arXiv Detail & Related papers (2022-08-30T00:39:06Z) - Continual Predictive Learning from Videos [100.27176974654559]
We study a new continual learning problem in the context of video prediction.
We propose the continual predictive learning (CPL) approach, which learns a mixture world model via predictive experience replay.
We construct two new benchmarks based on RoboNet and KTH, in which different tasks correspond to different physical robotic environments or human actions.
arXiv Detail & Related papers (2022-04-12T08:32:26Z) - Concept drift detection and adaptation for federated and continual
learning [55.41644538483948]
Smart devices can collect vast amounts of data from their environment.
This data is suitable for training machine learning models, which can significantly improve their behavior.
In this work, we present a new method, called Concept-Drift-Aware Federated Averaging.
arXiv Detail & Related papers (2021-05-27T17:01:58Z) - One to Many: Adaptive Instrument Segmentation via Meta Learning and
Dynamic Online Adaptation in Robotic Surgical Video [71.43912903508765]
MDAL is a dynamic online adaptive learning scheme for instrument segmentation in robot-assisted surgery.
It learns the general knowledge of instruments and the fast adaptation ability through the video-specific meta-learning paradigm.
It outperforms other state-of-the-art methods on two datasets.
arXiv Detail & Related papers (2021-03-24T05:02:18Z) - GymD2D: A Device-to-Device Underlay Cellular Offload Evaluation Platform [1.5863809575305414]
GymD2D is a framework for experimentation with physical layer resource allocation problems in device-to-device communication.
It allows users to simulate a variety of cellular offload scenarios and to extend its behaviour to meet their research needs.
We evaluate GymD2D with state-of-the-art deep reinforcement learning and demonstrate these algorithms provide significant efficiency improvements.
arXiv Detail & Related papers (2021-01-27T03:50:22Z) - Invariant Feature Learning for Sensor-based Human Activity Recognition [11.334750079923428]
We present an invariant feature learning framework (IFLF) that extracts common information shared across subjects and devices.
Experiments demonstrated that IFLF is effective in handling both subject and device diversion across popular open datasets and an in-house dataset.
arXiv Detail & Related papers (2020-12-14T21:56:17Z) - Never Stop Learning: The Effectiveness of Fine-Tuning in Robotic
Reinforcement Learning [109.77163932886413]
We show how to adapt vision-based robotic manipulation policies to new variations by fine-tuning via off-policy reinforcement learning.
This adaptation uses less than 0.2% of the data necessary to learn the task from scratch.
We find that our approach of adapting pre-trained policies leads to substantial performance gains over the course of fine-tuning.
arXiv Detail & Related papers (2020-04-21T17:57:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.