A Symbolic Representation of Human Posture for Interpretable Learning
and Reasoning
- URL: http://arxiv.org/abs/2210.08998v1
- Date: Mon, 17 Oct 2022 12:22:13 GMT
- Title: A Symbolic Representation of Human Posture for Interpretable Learning
and Reasoning
- Authors: Richard G. Freedman, Joseph B. Mueller, Jack Ladwig, Steven Johnston,
Helen Wauck, Ruta Wheelock, Hayley Borck
- Abstract summary: We introduce a qualitative spatial reasoning approach that describes the human posture in terms that are more familiar to people.
This paper explores the derivation of our symbolic representation at two levels of detail and its preliminary use as features for interpretable activity recognition.
- Score: 2.678461526933908
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Robots that interact with humans in a physical space or application need to
think about the person's posture, which typically comes from visual sensors
like cameras and infra-red. Artificial intelligence and machine learning
algorithms use information from these sensors either directly or after some
level of symbolic abstraction, and the latter usually partitions the range of
observed values to discretize the continuous signal data. Although these
representations have been effective in a variety of algorithms with respect to
accuracy and task completion, the underlying models are rarely interpretable,
which also makes their outputs more difficult to explain to people who request
them. Instead of focusing on the possible sensor values that are familiar to a
machine, we introduce a qualitative spatial reasoning approach that describes
the human posture in terms that are more familiar to people. This paper
explores the derivation of our symbolic representation at two levels of detail
and its preliminary use as features for interpretable activity recognition.
Related papers
- Adaptive Language-Guided Abstraction from Contrastive Explanations [53.48583372522492]
It is necessary to determine which features of the environment are relevant before determining how these features should be used to compute reward.
End-to-end methods for joint feature and reward learning often yield brittle reward functions that are sensitive to spurious state features.
This paper describes a method named ALGAE which alternates between using language models to iteratively identify human-meaningful features.
arXiv Detail & Related papers (2024-09-12T16:51:58Z) - Modeling User Preferences via Brain-Computer Interfacing [54.3727087164445]
We use Brain-Computer Interfacing technology to infer users' preferences, their attentional correlates towards visual content, and their associations with affective experience.
We link these to relevant applications, such as information retrieval, personalized steering of generative models, and crowdsourcing population estimates of affective experiences.
arXiv Detail & Related papers (2024-05-15T20:41:46Z) - Emotion Recognition from the perspective of Activity Recognition [0.0]
Appraising human emotional states, behaviors, and reactions displayed in real-world settings can be accomplished using latent continuous dimensions.
For emotion recognition systems to be deployed and integrated into real-world mobile and computing devices, we need to consider data collected in the world.
We propose a novel three-stream end-to-end deep learning regression pipeline with an attention mechanism.
arXiv Detail & Related papers (2024-03-24T18:53:57Z) - Human-oriented Representation Learning for Robotic Manipulation [64.59499047836637]
Humans inherently possess generalizable visual representations that empower them to efficiently explore and interact with the environments in manipulation tasks.
We formalize this idea through the lens of human-oriented multi-task fine-tuning on top of pre-trained visual encoders.
Our Task Fusion Decoder consistently improves the representation of three state-of-the-art visual encoders for downstream manipulation policy-learning.
arXiv Detail & Related papers (2023-10-04T17:59:38Z) - Towards Learning Discrete Representations via Self-Supervision for
Wearables-Based Human Activity Recognition [7.086647707011785]
Human activity recognition (HAR) in wearable computing is typically based on direct processing of sensor data.
Recent advancements in Vector Quantization (VQ) to wearables applications enables us to directly learn a mapping between short spans of sensor data and a codebook of vectors.
This work presents a proof-of-concept for demonstrating how effective discrete representations can be derived.
arXiv Detail & Related papers (2023-06-01T19:49:43Z) - Preserving Privacy in Human-Motion Affect Recognition [4.753703852165805]
This work evaluates the effectiveness of existing methods at recognising emotions using both 3D temporal joint signals and manually extracted features.
We propose a cross-subject transfer learning technique for training a multi-encoder autoencoder deep neural network to learn disentangled latent representations of human motion features.
arXiv Detail & Related papers (2021-05-09T15:26:21Z) - Matching Representations of Explainable Artificial Intelligence and Eye
Gaze for Human-Machine Interaction [0.7742297876120561]
Rapid non-verbal communication of task-based stimuli is a challenge in human-machine teaming.
In this work, we examine the correlations between visual heatmap explanations of a neural network trained to predict driving behavior and eye gaze heatmaps of human drivers.
arXiv Detail & Related papers (2021-01-30T07:42:56Z) - Task-relevant Representation Learning for Networked Robotic Perception [74.0215744125845]
This paper presents an algorithm to learn task-relevant representations of sensory data that are co-designed with a pre-trained robotic perception model's ultimate objective.
Our algorithm aggressively compresses robotic sensory data by up to 11x more than competing methods.
arXiv Detail & Related papers (2020-11-06T07:39:08Z) - DRG: Dual Relation Graph for Human-Object Interaction Detection [65.50707710054141]
We tackle the challenging problem of human-object interaction (HOI) detection.
Existing methods either recognize the interaction of each human-object pair in isolation or perform joint inference based on complex appearance-based features.
In this paper, we leverage an abstract spatial-semantic representation to describe each human-object pair and aggregate the contextual information of the scene via a dual relation graph.
arXiv Detail & Related papers (2020-08-26T17:59:40Z) - Human Activity Recognition from Wearable Sensor Data Using
Self-Attention [2.9023633922848586]
We present a self-attention based neural network model for activity recognition from body-worn sensor data.
We performed experiments on four popular publicly available HAR datasets: PAMAP2, Opportunity, Skoda and USC-HAD.
Our model achieve significant performance improvement over recent state-of-the-art models in both benchmark test subjects and Leave-one-out-subject evaluation.
arXiv Detail & Related papers (2020-03-17T14:16:57Z) - Continuous Emotion Recognition via Deep Convolutional Autoencoder and
Support Vector Regressor [70.2226417364135]
It is crucial that the machine should be able to recognize the emotional state of the user with high accuracy.
Deep neural networks have been used with great success in recognizing emotions.
We present a new model for continuous emotion recognition based on facial expression recognition.
arXiv Detail & Related papers (2020-01-31T17:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.