Efficient Deep Clustering of Human Activities and How to Improve
Evaluation
- URL: http://arxiv.org/abs/2209.08335v1
- Date: Sat, 17 Sep 2022 14:12:42 GMT
- Title: Efficient Deep Clustering of Human Activities and How to Improve
Evaluation
- Authors: Louis Mahon and Thomas Lukasiewicz
- Abstract summary: We present a new deep clustering model for human activity re-cog-ni-tion (HAR)
In this paper, we highlight several distinct problems with how deep HAR clustering models are evaluated.
We then discuss solutions to these problems, and suggest standard evaluation settings for future deep HAR clustering models.
- Score: 53.08810276824894
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There has been much recent research on human activity re\-cog\-ni\-tion
(HAR), due to the proliferation of wearable sensors in watches and phones, and
the advances of deep learning methods, which avoid the need to manually extract
features from raw sensor signals. A significant disadvantage of deep learning
applied to HAR is the need for manually labelled training data, which is
especially difficult to obtain for HAR datasets. Progress is starting to be
made in the unsupervised setting, in the form of deep HAR clustering models,
which can assign labels to data without having been given any labels to train
on, but there are problems with evaluating deep HAR clustering models, which
makes assessing the field and devising new methods difficult. In this paper, we
highlight several distinct problems with how deep HAR clustering models are
evaluated, describing these problems in detail and conducting careful
experiments to explicate the effect that they can have on results. We then
discuss solutions to these problems, and suggest standard evaluation settings
for future deep HAR clustering models. Additionally, we present a new deep
clustering model for HAR. When tested under our proposed settings, our model
performs better than (or on par with) existing models, while also being more
efficient and better able to scale to more complex datasets by avoiding the
need for an autoencoder.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - Automatic Dataset Construction (ADC): Sample Collection, Data Curation, and Beyond [38.89457061559469]
We propose an innovative methodology that automates dataset creation with negligible cost and high efficiency.
We provide open-source software that incorporates existing methods for label error detection, robust learning under noisy and biased data.
We design three benchmark datasets focused on label noise detection, label noise learning, and class-imbalanced learning.
arXiv Detail & Related papers (2024-08-21T04:45:12Z) - Deep Learning-Driven Approach for Handwritten Chinese Character Classification [0.0]
Handwritten character recognition is a challenging problem for machine learning researchers.
With numerous unique character classes present, some data, such as Logographic Scripts or Sino-Korean character sequences, bring new complications to the HCR problem.
This paper proposes a highly scalable approach for detailed character image classification by introducing the model architecture, data preprocessing steps, and testing design instructions.
arXiv Detail & Related papers (2024-01-30T15:29:32Z) - Zero-shot Retrieval: Augmenting Pre-trained Models with Search Engines [83.65380507372483]
Large pre-trained models can dramatically reduce the amount of task-specific data required to solve a problem, but they often fail to capture domain-specific nuances out of the box.
This paper shows how to leverage recent advances in NLP and multi-modal learning to augment a pre-trained model with search engine retrieval.
arXiv Detail & Related papers (2023-11-29T05:33:28Z) - Improving a Named Entity Recognizer Trained on Noisy Data with a Few
Clean Instances [55.37242480995541]
We propose to denoise noisy NER data with guidance from a small set of clean instances.
Along with the main NER model we train a discriminator model and use its outputs to recalibrate the sample weights.
Results on public crowdsourcing and distant supervision datasets show that the proposed method can consistently improve performance with a small guidance set.
arXiv Detail & Related papers (2023-10-25T17:23:37Z) - randomHAR: Improving Ensemble Deep Learners for Human Activity
Recognition with Sensor Selection and Reinforcement Learning [4.5830802802139585]
The general idea behind randomHAR is training a series of deep learning models with the same architecture on randomly selected sensor data.
In contrast to existing work, this approach optimize the ensemble process rather than the architecture of the constituent models.
The experiment demonstrates that the proposed approach outperforms the state-of-the-art method, ensembleLSTM.
arXiv Detail & Related papers (2023-07-15T10:51:03Z) - STAR: Boosting Low-Resource Information Extraction by Structure-to-Text
Data Generation with Large Language Models [56.27786433792638]
STAR is a data generation method that leverages Large Language Models (LLMs) to synthesize data instances.
We design fine-grained step-by-step instructions to obtain the initial data instances.
Our experiments show that the data generated by STAR significantly improve the performance of low-resource event extraction and relation extraction tasks.
arXiv Detail & Related papers (2023-05-24T12:15:19Z) - Discrete Key-Value Bottleneck [95.61236311369821]
Deep neural networks perform well on classification tasks where data streams are i.i.d. and labeled data is abundant.
One powerful approach that has addressed this challenge involves pre-training of large encoders on volumes of readily available data, followed by task-specific tuning.
Given a new task, however, updating the weights of these encoders is challenging as a large number of weights needs to be fine-tuned, and as a result, they forget information about the previous tasks.
We propose a model architecture to address this issue, building upon a discrete bottleneck containing pairs of separate and learnable key-value codes.
arXiv Detail & Related papers (2022-07-22T17:52:30Z) - Learning-to-Learn Personalised Human Activity Recognition Models [1.5087842661221904]
We present a meta-learning methodology for learning to learn personalised HAR models for HAR.
We introduce two algorithms, Personalised MAML and Personalised Relation Networks inspired by existing Meta-Learning algorithms.
A comparative study shows significant performance improvements against the state-of-the-art Deep Learning algorithms and the Few-shot Meta-Learning algorithms in multiple HAR domains.
arXiv Detail & Related papers (2020-06-12T21:11:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.