REFORM: Recognizing F-formations for Social Robots
- URL: http://arxiv.org/abs/2008.07668v1
- Date: Mon, 17 Aug 2020 23:32:05 GMT
- Title: REFORM: Recognizing F-formations for Social Robots
- Authors: Hooman Hedayati, Annika Muehlbradt, Daniel J. Szafir, Sean Andrist
- Abstract summary: We introduce REFORM, a data-driven approach for detecting F-formations given human and agent positions and orientations.
We find that REFORM yielded improved accuracy over a state-of-the-art F-formation detection algorithm.
- Score: 4.833815605196964
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recognizing and understanding conversational groups, or F-formations, is a
critical task for situated agents designed to interact with humans.
F-formations contain complex structures and dynamics, yet are used intuitively
by people in everyday face-to-face conversations. Prior research exploring ways
of identifying F-formations has largely relied on heuristic algorithms that may
not capture the rich dynamic behaviors employed by humans. We introduce REFORM
(REcognize F-FORmations with Machine learning), a data-driven approach for
detecting F-formations given human and agent positions and orientations. REFORM
decomposes the scene into all possible pairs and then reconstructs F-formations
with a voting-based scheme. We evaluated our approach across three datasets:
the SALSA dataset, a newly collected human-only dataset, and a new set of acted
human-robot scenarios, and found that REFORM yielded improved accuracy over a
state-of-the-art F-formation detection algorithm. We also introduce symmetry
and tightness as quantitative measures to characterize F-formations.
Supplementary video: https://youtu.be/Fp7ETdkKvdA , Dataset available at:
github.com/cu-ironlab/Babble
Related papers
- Belief Revision: The Adaptability of Large Language Models Reasoning [63.0281286287648]
We introduce Belief-R, a new dataset designed to test LMs' belief revision ability when presented with new evidence.
Inspired by how humans suppress prior inferences, this task assesses LMs within the newly proposed delta reasoning framework.
We evaluate $sim$30 LMs across diverse prompting strategies and found that LMs generally struggle to appropriately revise their beliefs in response to new information.
arXiv Detail & Related papers (2024-06-28T09:09:36Z) - PolyFit: A Peg-in-hole Assembly Framework for Unseen Polygon Shapes via
Sim-to-real Adaptation [4.875369637162596]
PolyFit is a supervised learning framework designed for 5-DoF peg-in-hole assembly.
It utilizes F/T data for accurate extrinsic pose estimation and adjusts the peg pose to rectify misalignments.
It achieves impressive peg-in-hole success rates of 97.3% and 96.3% for seen and unseen shapes in simulations.
arXiv Detail & Related papers (2023-12-05T06:28:33Z) - Generalized Face Forgery Detection via Adaptive Learning for Pre-trained Vision Transformer [54.32283739486781]
We present a textbfForgery-aware textbfAdaptive textbfVision textbfTransformer (FA-ViT) under the adaptive learning paradigm.
FA-ViT achieves 93.83% and 78.32% AUC scores on Celeb-DF and DFDC datasets in the cross-dataset evaluation.
arXiv Detail & Related papers (2023-09-20T06:51:11Z) - A Comprehensive Survey on Pretrained Foundation Models: A History from
BERT to ChatGPT [0.0]
Pretrained Foundation Models (PFMs) are regarded as the foundation for various downstream tasks with different data modalities.
This study provides a comprehensive review of recent research advancements, challenges, and opportunities for PFMs in text, image, graph, as well as other data modalities.
arXiv Detail & Related papers (2023-02-18T20:51:09Z) - F2SD: A dataset for end-to-end group detection algorithms [3.3117512968892355]
We develop a large-scale dataset of simulated images for F-formation detection, called F-formation Simulation dataset (F2SD)
F2SD contains nearly 60,000 images simulated from GTA-5, with bounding boxes and orientation information on images.
It is challenging to construct such a large-scale simulated dataset while keeping it realistic.
arXiv Detail & Related papers (2022-11-20T15:42:22Z) - Predicting Loose-Fitting Garment Deformations Using Bone-Driven Motion
Networks [63.596602299263935]
We present a learning algorithm that uses bone-driven motion networks to predict the deformation of loose-fitting garment meshes at interactive rates.
We show that our method outperforms state-of-the-art methods in terms of prediction accuracy of mesh deformations by about 20% in RMSE and 10% in Hausdorff distance and STED.
arXiv Detail & Related papers (2022-05-03T07:54:39Z) - Rethinking Architecture Design for Tackling Data Heterogeneity in
Federated Learning [53.73083199055093]
We show that attention-based architectures (e.g., Transformers) are fairly robust to distribution shifts.
Our experiments show that replacing convolutional networks with Transformers can greatly reduce catastrophic forgetting of previous devices.
arXiv Detail & Related papers (2021-06-10T21:04:18Z) - Locally Aware Piecewise Transformation Fields for 3D Human Mesh
Registration [67.69257782645789]
We propose piecewise transformation fields that learn 3D translation vectors to map any query point in posed space to its correspond position in rest-pose space.
We show that fitting parametric models with poses by our network results in much better registration quality, especially for extreme poses.
arXiv Detail & Related papers (2021-04-16T15:16:09Z) - Let me join you! Real-time F-formation recognition by a socially aware
robot [2.8101673772585745]
This paper presents a novel architecture to detect social groups in real-time from a continuous image stream of an ego-vision camera.
We detect F-formations in social gatherings such as meetings, discussions, etc. and predict the robot's approach angle if it wants to join the social group.
We also detect outliers, i.e., the persons who are not part of the group under consideration.
arXiv Detail & Related papers (2020-08-23T17:46:08Z) - Joint Inference of States, Robot Knowledge, and Human (False-)Beliefs [90.20235972293801]
Aiming to understand how human (false-temporal)-belief-a core socio-cognitive ability unify-would affect human interactions with robots, this paper proposes to adopt a graphical model to the representation of object states, robot knowledge, and human (false-)beliefs.
An inference algorithm is derived to fuse individual pg from all robots across multi-views into a joint pg, which affords more effective reasoning inference capability to overcome the errors originated from a single view.
arXiv Detail & Related papers (2020-04-25T23:02:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.