A Multi-Task Deep Learning Approach for Sensor-based Human Activity
Recognition and Segmentation
- URL: http://arxiv.org/abs/2303.11100v1
- Date: Mon, 20 Mar 2023 13:34:28 GMT
- Title: A Multi-Task Deep Learning Approach for Sensor-based Human Activity
Recognition and Segmentation
- Authors: Furong Duan, Tao Zhu, Jinqiang Wang, Liming Chen, Huansheng Ning,
Yaping Wan
- Abstract summary: We propose a new deep neural network to solve the two tasks simultaneously.
The proposed network adopts selective convolution and features multiscale windows to segment activities of long or short time durations.
Our proposed method outperforms the state-of-the-art methods both for activity recognition and segmentation.
- Score: 4.987833356397567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sensor-based human activity segmentation and recognition are two important
and challenging problems in many real-world applications and they have drawn
increasing attention from the deep learning community in recent years. Most of
the existing deep learning works were designed based on pre-segmented sensor
streams and they have treated activity segmentation and recognition as two
separate tasks. In practice, performing data stream segmentation is very
challenging. We believe that both activity segmentation and recognition may
convey unique information which can complement each other to improve the
performance of the two tasks. In this paper, we firstly proposes a new
multitask deep neural network to solve the two tasks simultaneously. The
proposed neural network adopts selective convolution and features multiscale
windows to segment activities of long or short time durations. First, multiple
windows of different scales are generated to center on each unit of the feature
sequence. Then, the model is trained to predict, for each window, the activity
class and the offset to the true activity boundaries. Finally, overlapping
windows are filtered out by non-maximum suppression, and adjacent windows of
the same activity are concatenated to complete the segmentation task. Extensive
experiments were conducted on eight popular benchmarking datasets, and the
results show that our proposed method outperforms the state-of-the-art methods
both for activity recognition and segmentation.
Related papers
- Frequency-based Matcher for Long-tailed Semantic Segmentation [22.199174076366003]
We focus on a relatively under-explored task setting, long-tailed semantic segmentation (LTSS)
We propose a dual-metric evaluation system and construct the LTSS benchmark to demonstrate the performance of semantic segmentation methods and long-tailed solutions.
We also propose a transformer-based algorithm to improve LTSS, frequency-based matcher, which solves the oversuppression problem by one-to-many matching.
arXiv Detail & Related papers (2024-06-06T09:57:56Z) - Auxiliary Tasks Enhanced Dual-affinity Learning for Weakly Supervised
Semantic Segmentation [79.05949524349005]
We propose AuxSegNet+, a weakly supervised auxiliary learning framework to explore the rich information from saliency maps.
We also propose a cross-task affinity learning mechanism to learn pixel-level affinities from the saliency and segmentation feature maps.
arXiv Detail & Related papers (2024-03-02T10:03:21Z) - A Dynamic Feature Interaction Framework for Multi-task Visual Perception [100.98434079696268]
We devise an efficient unified framework to solve multiple common perception tasks.
These tasks include instance segmentation, semantic segmentation, monocular 3D detection, and depth estimation.
Our proposed framework, termed D2BNet, demonstrates a unique approach to parameter-efficient predictions for multi-task perception.
arXiv Detail & Related papers (2023-06-08T09:24:46Z) - Remote Task-oriented Grasp Area Teaching By Non-Experts through
Interactive Segmentation and Few-Shot Learning [0.0]
A robot must be able to discriminate between different grasping styles depending on the prospective manipulation task.
We propose a novel two-step framework towards this aim.
We receive grasp area demonstrations for a new task via interactive segmentation.
We learn from these few demonstrations to estimate the required grasp area on an unseen scene for the given task.
arXiv Detail & Related papers (2023-03-17T18:09:01Z) - Visual Exemplar Driven Task-Prompting for Unified Perception in
Autonomous Driving [100.3848723827869]
We present an effective multi-task framework, VE-Prompt, which introduces visual exemplars via task-specific prompting.
Specifically, we generate visual exemplars based on bounding boxes and color-based markers, which provide accurate visual appearances of target categories.
We bridge transformer-based encoders and convolutional layers for efficient and accurate unified perception in autonomous driving.
arXiv Detail & Related papers (2023-03-03T08:54:06Z) - Multi-level Contrast Network for Wearables-based Joint Activity
Segmentation and Recognition [10.828099015828693]
Human activity recognition (HAR) with wearables is promising research that can be widely adopted in many smart healthcare applications.
Most HAR algorithms are susceptible to the multi-class windows problem that is essential yet rarely exploited.
We introduce the segmentation technology into HAR, yielding joint activity segmentation and recognition.
arXiv Detail & Related papers (2022-08-16T05:39:02Z) - Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation [88.49669148290306]
We propose a novel weakly supervised multi-task framework called AuxSegNet to leverage saliency detection and multi-label image classification as auxiliary tasks.
Inspired by their similar structured semantics, we also propose to learn a cross-task global pixel-level affinity map from the saliency and segmentation representations.
The learned cross-task affinity can be used to refine saliency predictions and propagate CAM maps to provide improved pseudo labels for both tasks.
arXiv Detail & Related papers (2021-07-25T11:39:58Z) - Distribution Alignment: A Unified Framework for Long-tail Visual
Recognition [52.36728157779307]
We propose a unified distribution alignment strategy for long-tail visual recognition.
We then introduce a generalized re-weight method in the two-stage learning to balance the class prior.
Our approach achieves the state-of-the-art results across all four recognition tasks with a simple and unified framework.
arXiv Detail & Related papers (2021-03-30T14:09:53Z) - A Tree-structure Convolutional Neural Network for Temporal Features
Exaction on Sensor-based Multi-resident Activity Recognition [4.619245607612873]
We propose an end-to-end Tree-Structure Convolutional neural network based framework for Multi-Resident Activity Recognition (TSC-MRAR)
First, we treat each sample as an event and obtain the current event embedding through the previous sensor readings in the sliding window.
Then, in order to automatically generate the temporal features, a tree-structure network is designed to derive the temporal dependence of nearby readings.
arXiv Detail & Related papers (2020-11-05T14:31:00Z) - Sequential Weakly Labeled Multi-Activity Localization and Recognition on
Wearable Sensors using Recurrent Attention Networks [13.64024154785943]
We propose a recurrent attention network (RAN) to handle sequential weakly labeled multi-activity recognition and location tasks.
Our RAN model can simultaneously infer multi-activity types from the coarse-grained sequential weak labels.
It will greatly reduce the burden of manual labeling.
arXiv Detail & Related papers (2020-04-13T04:57:09Z) - ZSTAD: Zero-Shot Temporal Activity Detection [107.63759089583382]
We propose a novel task setting called zero-shot temporal activity detection (ZSTAD), where activities that have never been seen in training can still be detected.
We design an end-to-end deep network based on R-C3D as the architecture for this solution.
Experiments on both the THUMOS14 and the Charades datasets show promising performance in terms of detecting unseen activities.
arXiv Detail & Related papers (2020-03-12T02:40:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.