Action similarity judgment based on kinematic primitives
- URL: http://arxiv.org/abs/2008.13176v1
- Date: Sun, 30 Aug 2020 13:58:47 GMT
- Title: Action similarity judgment based on kinematic primitives
- Authors: Vipul Nair, Paul Hemeren, Alessia Vignolo, Nicoletta Noceti, Elena
Nicora, Alessandra Sciutti, Francesco Rea, Erik Billing, Francesca Odone and
Giulio Sandini
- Abstract summary: We investigate to which extent a computational model based on kinematics can determine action similarity.
The chosen model has its roots in developmental robotics and performs action classification based on learned kinematic primitives.
The results show that both the model and human performance are highly accurate in an action similarity task based on kinematic-level features.
- Score: 48.99831733355487
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding which features humans rely on -- in visually recognizing action
similarity is a crucial step towards a clearer picture of human action
perception from a learning and developmental perspective. In the present work,
we investigate to which extent a computational model based on kinematics can
determine action similarity and how its performance relates to human similarity
judgments of the same actions. To this aim, twelve participants perform an
action similarity task, and their performances are compared to that of a
computational model solving the same task. The chosen model has its roots in
developmental robotics and performs action classification based on learned
kinematic primitives. The comparative experiment results show that both the
model and human participants can reliably identify whether two actions are the
same or not. However, the model produces more false hits and has a greater
selection bias than human participants. A possible reason for this is the
particular sensitivity of the model towards kinematic primitives of the
presented actions. In a second experiment, human participants' performance on
an action identification task indicated that they relied solely on kinematic
information rather than on action semantics. The results show that both the
model and human performance are highly accurate in an action similarity task
based on kinematic-level features, which can provide an essential basis for
classifying human actions.
Related papers
- Are Visual-Language Models Effective in Action Recognition? A Comparative Study [22.97135293252601]
This paper provides a large-scale study and insight on state-of-the-art vision foundation models.
It compares their transfer ability onto zero-shot and frame-wise action recognition tasks.
Experiments are conducted on recent fine-grained, human-centric action recognition datasets.
arXiv Detail & Related papers (2024-10-22T16:28:21Z) - Evaluating Multiview Object Consistency in Humans and Image Models [68.36073530804296]
We leverage an experimental design from the cognitive sciences which requires zero-shot visual inferences about object shape.
We collect 35K trials of behavioral data from over 500 participants.
We then evaluate the performance of common vision models.
arXiv Detail & Related papers (2024-09-09T17:59:13Z) - A Training Method For VideoPose3D With Ideology of Action Recognition [0.9949781365631559]
This research shows a faster and more flexible training method for VideoPose3D based on action recognition.
It can handle both action-oriented and common pose-estimation problems.
arXiv Detail & Related papers (2022-06-13T19:25:27Z) - Skeleton-Based Mutually Assisted Interacted Object Localization and
Human Action Recognition [111.87412719773889]
We propose a joint learning framework for "interacted object localization" and "human action recognition" based on skeleton data.
Our method achieves the best or competitive performance with the state-of-the-art methods for human action recognition.
arXiv Detail & Related papers (2021-10-28T10:09:34Z) - Spatio-Temporal Human Action Recognition Modelwith Flexible-interval
Sampling and Normalization [0.0]
We propose a human action system for Red-Green-Blue(RGB) video input with our own designed module.
We build a novel dataset with a similar background and discriminative actions for both human keypoint prediction and behavior recognition.
Experimental results demonstrate the effectiveness of the proposed model on our own human behavior recognition dataset and some public datasets.
arXiv Detail & Related papers (2021-08-12T10:02:20Z) - STAR: Sparse Transformer-based Action Recognition [61.490243467748314]
This work proposes a novel skeleton-based human action recognition model with sparse attention on the spatial dimension and segmented linear attention on the temporal dimension of data.
Experiments show that our model can achieve comparable performance while utilizing much less trainable parameters and achieve high speed in training and inference.
arXiv Detail & Related papers (2021-07-15T02:53:11Z) - A robot that counts like a child: a developmental model of counting and
pointing [69.26619423111092]
A novel neuro-robotics model capable of counting real items is introduced.
The model allows us to investigate the interaction between embodiment and numerical cognition.
The trained model is able to count a set of items and at the same time points to them.
arXiv Detail & Related papers (2020-08-05T21:06:27Z) - Few-shot Visual Reasoning with Meta-analogical Contrastive Learning [141.2562447971]
We propose to solve a few-shot (or low-shot) visual reasoning problem, by resorting to analogical reasoning.
We extract structural relationships between elements in both domains, and enforce them to be as similar as possible with analogical learning.
We validate our method on RAVEN dataset, on which it outperforms state-of-the-art method, with larger gains when the training data is scarce.
arXiv Detail & Related papers (2020-07-23T14:00:34Z) - Learning intuitive physics and one-shot imitation using
state-action-prediction self-organizing maps [0.0]
Humans learn by exploration and imitation, build causal models of the world, and use both to flexibly solve new tasks.
We suggest a simple but effective unsupervised model which develops such characteristics.
We demonstrate its performance on a set of several related, but different one-shot imitation tasks, which the agent flexibly solves in an active inference style.
arXiv Detail & Related papers (2020-07-03T12:29:11Z) - Simultaneous Learning from Human Pose and Object Cues for Real-Time
Activity Recognition [11.290467061493189]
We propose a novel approach to real-time human activity recognition, through simultaneously learning from observations of both human poses and objects involved in the human activity.
Our method outperforms previous methods and obtains real-time performance for human activity recognition with a processing speed of 104 Hz.
arXiv Detail & Related papers (2020-03-26T22:04:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.