Applications of human activity recognition in industrial processes --
Synergy of human and technology
- URL: http://arxiv.org/abs/2212.02266v1
- Date: Mon, 5 Dec 2022 13:45:45 GMT
- Title: Applications of human activity recognition in industrial processes --
Synergy of human and technology
- Authors: Friedrich Niemann, Christopher Reining, H\"ulya Bas and Sven Franke
- Abstract summary: We introduce ongoing research on human activity recognition in intralogistics.
We show how semantic attributes can be used to describe human activities flexibly.
We present a concept based on a cyber-physical twin that can reduce the effort and time necessary to create a training dataset.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human-technology collaboration relies on verbal and non-verbal communication.
Machines must be able to detect and understand the movements of humans to
facilitate non-verbal communication. In this article, we introduce ongoing
research on human activity recognition in intralogistics, and show how it can
be applied in industrial settings. We show how semantic attributes can be used
to describe human activities flexibly and how context informantion increases
the performance of classifiers to recognise them automatically. Beyond that, we
present a concept based on a cyber-physical twin that can reduce the effort and
time necessary to create a training dataset for human activity recognition. In
the future, it will be possible to train a classifier solely with realistic
simulation data, while maintaining or even increasing the classification
performance.
Related papers
- Visual-Geometric Collaborative Guidance for Affordance Learning [63.038406948791454]
We propose a visual-geometric collaborative guided affordance learning network that incorporates visual and geometric cues.
Our method outperforms the representative models regarding objective metrics and visual quality.
arXiv Detail & Related papers (2024-10-15T07:35:51Z) - Multimodal Fusion with LLMs for Engagement Prediction in Natural Conversation [70.52558242336988]
We focus on predicting engagement in dyadic interactions by scrutinizing verbal and non-verbal cues, aiming to detect signs of disinterest or confusion.
In this work, we collect a dataset featuring 34 participants engaged in casual dyadic conversations, each providing self-reported engagement ratings at the end of each conversation.
We introduce a novel fusion strategy using Large Language Models (LLMs) to integrate multiple behavior modalities into a multimodal transcript''
arXiv Detail & Related papers (2024-09-13T18:28:12Z) - Real-time Addressee Estimation: Deployment of a Deep-Learning Model on
the iCub Robot [52.277579221741746]
Addressee Estimation is a skill essential for social robots to interact smoothly with humans.
Inspired by human perceptual skills, a deep-learning model for Addressee Estimation is designed, trained, and deployed on an iCub robot.
The study presents the procedure of such implementation and the performance of the model deployed in real-time human-robot interaction.
arXiv Detail & Related papers (2023-11-09T13:01:21Z) - CHARM: A Hierarchical Deep Learning Model for Classification of Complex
Human Activities Using Motion Sensors [0.9594432031144714]
CHARM is a hierarchical deep learning model for classification of complex human activities using motion sensors.
It outperforms state-of-the-art supervised learning approaches for high-level activity recognition in terms of average accuracy and F1 scores.
The ability to learn low-level user activities when trained using only high-level activity labels may pave the way to semi-supervised learning of HAR tasks.
arXiv Detail & Related papers (2022-07-16T01:36:54Z) - BOSS: A Benchmark for Human Belief Prediction in Object-context
Scenarios [14.23697277904244]
This paper uses the combined knowledge of Theory of Mind (ToM) and Object-Context Relations to investigate methods for enhancing collaboration between humans and autonomous systems.
We propose a novel and challenging multimodal video dataset for assessing the capability of artificial intelligence (AI) systems in predicting human belief states in an object-context scenario.
arXiv Detail & Related papers (2022-06-21T18:29:17Z) - Data-driven emotional body language generation for social robotics [58.88028813371423]
In social robotics, endowing humanoid robots with the ability to generate bodily expressions of affect can improve human-robot interaction and collaboration.
We implement a deep learning data-driven framework that learns from a few hand-designed robotic bodily expressions.
The evaluation study found that the anthropomorphism and animacy of the generated expressions are not perceived differently from the hand-designed ones.
arXiv Detail & Related papers (2022-05-02T09:21:39Z) - Cognitive architecture aided by working-memory for self-supervised
multi-modal humans recognition [54.749127627191655]
The ability to recognize human partners is an important social skill to build personalized and long-term human-robot interactions.
Deep learning networks have achieved state-of-the-art results and demonstrated to be suitable tools to address such a task.
One solution is to make robots learn from their first-hand sensory data with self-supervision.
arXiv Detail & Related papers (2021-03-16T13:50:24Z) - Human Activity Recognition Using Multichannel Convolutional Neural
Network [0.0]
Human Activity Recognition (HAR) simply refers to the capacity of a machine to perceive human actions.
This paper describes a supervised learning method that can distinguish human actions based on data collected from practical human movements.
The model was tested on the UCI HAR dataset, which resulted in a 95.25% classification accuracy.
arXiv Detail & Related papers (2021-01-17T16:48:17Z) - Simultaneous Learning from Human Pose and Object Cues for Real-Time
Activity Recognition [11.290467061493189]
We propose a novel approach to real-time human activity recognition, through simultaneously learning from observations of both human poses and objects involved in the human activity.
Our method outperforms previous methods and obtains real-time performance for human activity recognition with a processing speed of 104 Hz.
arXiv Detail & Related papers (2020-03-26T22:04:37Z) - Continuous Emotion Recognition via Deep Convolutional Autoencoder and
Support Vector Regressor [70.2226417364135]
It is crucial that the machine should be able to recognize the emotional state of the user with high accuracy.
Deep neural networks have been used with great success in recognizing emotions.
We present a new model for continuous emotion recognition based on facial expression recognition.
arXiv Detail & Related papers (2020-01-31T17:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.