hear-your-action: human action recognition by ultrasound active sensing
- URL: http://arxiv.org/abs/2309.08087v1
- Date: Fri, 15 Sep 2023 01:00:55 GMT
- Title: hear-your-action: human action recognition by ultrasound active sensing
- Authors: Risako Tanigawa, Yasunori Ishii
- Abstract summary: Action recognition is a key technology for many industrial applications.
Privacy issues prevent widespread usage due to the inclusion of private information.
We propose a privacy-preserving action recognition by ultrasound active sensing.
- Score: 3.0277213703725767
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Action recognition is a key technology for many industrial applications.
Methods using visual information such as images are very popular. However,
privacy issues prevent widespread usage due to the inclusion of private
information, such as visible faces and scene backgrounds, which are not
necessary for recognizing user action. In this paper, we propose a
privacy-preserving action recognition by ultrasound active sensing. As action
recognition from ultrasound active sensing in a non-invasive manner is not well
investigated, we create a new dataset for action recognition and conduct a
comparison of features for classification. We calculated feature values by
focusing on the temporal variation of the amplitude of ultrasound reflected
waves and performed classification using a support vector machine and VGG for
eight fundamental action classes. We confirmed that our method achieved an
accuracy of 97.9% when trained and evaluated on the same person and in the same
environment. Additionally, our method achieved an accuracy of 89.5% even when
trained and evaluated on different people. We also report the analyses of
accuracies in various conditions and limitations.
Related papers
- A noisy elephant in the room: Is your out-of-distribution detector robust to label noise? [49.88894124047644]
We take a closer look at 20 state-of-the-art OOD detection methods.
We show that poor separation between incorrectly classified ID samples vs. OOD samples is an overlooked yet important limitation of existing methods.
arXiv Detail & Related papers (2024-04-02T09:40:22Z) - Emotion Recognition from the perspective of Activity Recognition [0.0]
Appraising human emotional states, behaviors, and reactions displayed in real-world settings can be accomplished using latent continuous dimensions.
For emotion recognition systems to be deployed and integrated into real-world mobile and computing devices, we need to consider data collected in the world.
We propose a novel three-stream end-to-end deep learning regression pipeline with an attention mechanism.
arXiv Detail & Related papers (2024-03-24T18:53:57Z) - Measuring Human Perception to Improve Open Set Recognition [4.124573231232705]
Human ability to recognize when an object belongs or does not belong to a particular vision task outperforms all open set recognition algorithms.
measured reaction time from human subjects can offer insight as to whether a class sample is prone to be confused with a different class.
New psychophysical loss function enforces consistency with human behavior in deep networks which exhibit variable reaction time for different images.
arXiv Detail & Related papers (2022-09-08T01:19:36Z) - Privacy-Preserving Face Recognition with Learnable Privacy Budgets in
Frequency Domain [77.8858706250075]
This paper proposes a privacy-preserving face recognition method using differential privacy in the frequency domain.
Our method performs very well with several classical face recognition test sets.
arXiv Detail & Related papers (2022-07-15T07:15:36Z) - Invisible-to-Visible: Privacy-Aware Human Instance Segmentation using
Airborne Ultrasound via Collaborative Learning Variational Autoencoder [8.21448246263952]
We propose a new task for human instance segmentation from invisible information, especially airborne ultrasound, for action recognition.
To perform instance segmentation from invisible information, we first convert sound waves to reflected sound directional images (sound images)
In inference, it is possible to obtain instance segmentation results only from sound images.
arXiv Detail & Related papers (2022-04-15T00:56:01Z) - Audio-Visual Person-of-Interest DeepFake Detection [77.04789677645682]
The aim of this work is to propose a deepfake detector that can cope with the wide variety of manipulation methods and scenarios encountered in the real world.
We leverage a contrastive learning paradigm to learn the moving-face and audio segment embeddings that are most discriminative for each identity.
Our method can detect both single-modality (audio-only, video-only) and multi-modality (audio-video) attacks, and is robust to low-quality or corrupted videos.
arXiv Detail & Related papers (2022-04-06T20:51:40Z) - Skeleton-Based Mutually Assisted Interacted Object Localization and
Human Action Recognition [111.87412719773889]
We propose a joint learning framework for "interacted object localization" and "human action recognition" based on skeleton data.
Our method achieves the best or competitive performance with the state-of-the-art methods for human action recognition.
arXiv Detail & Related papers (2021-10-28T10:09:34Z) - Spotting adversarial samples for speaker verification by neural vocoders [102.1486475058963]
We adopt neural vocoders to spot adversarial samples for automatic speaker verification (ASV)
We find that the difference between the ASV scores for the original and re-synthesize audio is a good indicator for discrimination between genuine and adversarial samples.
Our codes will be made open-source for future works to do comparison.
arXiv Detail & Related papers (2021-07-01T08:58:16Z) - A Grid-based Representation for Human Action Recognition [12.043574473965318]
Human action recognition (HAR) in videos is a fundamental research topic in computer vision.
We propose a novel method for action recognition that encodes efficiently the most discriminative appearance information of an action.
Our method is tested on several benchmark datasets demonstrating that our model can accurately recognize human actions.
arXiv Detail & Related papers (2020-10-17T18:25:00Z) - Cross-Task Transfer for Geotagged Audiovisual Aerial Scene Recognition [61.54648991466747]
We explore an audiovisual aerial scene recognition task using both images and sounds as input.
We show the benefit of exploiting the audio information for the aerial scene recognition.
arXiv Detail & Related papers (2020-05-18T04:14:16Z) - Sensor Data for Human Activity Recognition: Feature Representation and
Benchmarking [27.061240686613182]
The field of Human Activity Recognition (HAR) focuses on obtaining and analysing data captured from monitoring devices (e.g. sensors)
We address the issue of accurately recognising human activities using different Machine Learning (ML) techniques.
arXiv Detail & Related papers (2020-05-15T00:46:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.