Individual vs. Joint Perception: a Pragmatic Model of Pointing as
Communicative Smithian Helping
- URL: http://arxiv.org/abs/2106.02003v1
- Date: Thu, 3 Jun 2021 17:21:23 GMT
- Title: Individual vs. Joint Perception: a Pragmatic Model of Pointing as
Communicative Smithian Helping
- Authors: Kaiwen Jiang, Stephanie Stacy, Chuyu Wei, Adelpha Chan, Federico
Rossano, Yixin Zhu, Tao Gao
- Abstract summary: The simple gesture of pointing can greatly augment ones ability to comprehend states of the world based on observations.
We model an agents update to its belief of the world based on individual observations using a partially observable Markov decision process (POMDP)
On top of that, we model pointing as a communicative act between agents who have a mutual understanding that the pointed observation must be relevant and interpretable.
- Score: 16.671443846399836
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The simple gesture of pointing can greatly augment ones ability to comprehend
states of the world based on observations. It triggers additional inferences
relevant to ones task at hand. We model an agents update to its belief of the
world based on individual observations using a partially observable Markov
decision process (POMDP), a mainstream artificial intelligence (AI) model of
how to act rationally according to beliefs formed through observation. On top
of that, we model pointing as a communicative act between agents who have a
mutual understanding that the pointed observation must be relevant and
interpretable. Our model measures relevance by defining a Smithian Value of
Information (SVI) as the utility improvement of the POMDP agent before and
after receiving the pointing. We model that agents calculate SVI by using the
cognitive theory of Smithian helping as a principle of coordinating separate
beliefs for action prediction and action evaluation. We then import SVI into
rational speech act (RSA) as the utility function of an utterance. These lead
us to a pragmatic model of pointing allowing for contextually flexible
interpretations. We demonstrate the power of our Smithian pointing model by
extending the Wumpus world, a classic AI task where a hunter hunts a monster
with only partial observability of the world. We add another agent as a guide
who can only help by marking an observation already perceived by the hunter
with a pointing or not, without providing new observations or offering any
instrumental help. Our results show that this severely limited and overloaded
communication nevertheless significantly improves the hunters performance. The
advantage of pointing is indeed due to a computation of relevance based on
Smithian helping, as it disappears completely when the task is too difficult or
too easy for the guide to help.
Related papers
- Observation Interference in Partially Observable Assistance Games [34.53170543153206]
We study a model of the human-AI value alignment problem which allows the human and the AI assistant to have partial observations.
We show that sometimes an optimal assistant must take observation-interfering actions, even when the human is playing optimally.
We show that if the human acts according to the Boltzmann model of irrationality, this can create an incentive for the assistant to interfere with observations.
arXiv Detail & Related papers (2024-12-23T18:53:33Z) - Improving Zero-Shot ObjectNav with Generative Communication [60.84730028539513]
We propose a new method for improving zero-shot ObjectNav.
Our approach takes into account that the ground agent may have limited and sometimes obstructed view.
arXiv Detail & Related papers (2024-08-03T22:55:26Z) - Understanding Probe Behaviors through Variational Bounds of Mutual
Information [53.520525292756005]
We provide guidelines for linear probing by constructing a novel mathematical framework leveraging information theory.
First, we connect probing with the variational bounds of mutual information (MI) to relax the probe design, equating linear probing with fine-tuning.
We show that the intermediate representations can have the biggest MI estimate because of the tradeoff between better separability and decreasing MI.
arXiv Detail & Related papers (2023-12-15T18:38:18Z) - Sim-to-Real Causal Transfer: A Metric Learning Approach to
Causally-Aware Interaction Representations [62.48505112245388]
We take an in-depth look at the causal awareness of modern representations of agent interactions.
We show that recent representations are already partially resilient to perturbations of non-causal agents.
We propose a metric learning approach that regularizes latent representations with causal annotations.
arXiv Detail & Related papers (2023-12-07T18:57:03Z) - Post Hoc Explanations of Language Models Can Improve Language Models [43.2109029463221]
We present a novel framework, Amplifying Model Performance by Leveraging In-Context Learning with Post Hoc Explanations (AMPLIFY)
We leverage post hoc explanation methods which output attribution scores (explanations) capturing the influence of each of the input features on model predictions.
Our framework, AMPLIFY, leads to prediction accuracy improvements of about 10-25% over a wide range of tasks.
arXiv Detail & Related papers (2023-05-19T04:46:04Z) - Imitation from Observation With Bootstrapped Contrastive Learning [12.048166025000976]
Imitation from observation (IfO) is a learning paradigm that consists of training autonomous agents in a Markov Decision Process.
We present BootIfOL, an IfO algorithm that aims to learn a reward function that takes an agent trajectory and compares it to an expert.
We evaluate our approach on a variety of control tasks showing that we can train effective policies using a limited number of demonstrative trajectories.
arXiv Detail & Related papers (2023-02-13T17:32:17Z) - H-SAUR: Hypothesize, Simulate, Act, Update, and Repeat for Understanding
Object Articulations from Interactions [62.510951695174604]
"Hypothesize, Simulate, Act, Update, and Repeat" (H-SAUR) is a probabilistic generative framework that generates hypotheses about how objects articulate given input observations.
We show that the proposed model significantly outperforms the current state-of-the-art articulated object manipulation framework.
We further improve the test-time efficiency of H-SAUR by integrating a learned prior from learning-based vision models.
arXiv Detail & Related papers (2022-10-22T18:39:33Z) - Beyond Tracking: Using Deep Learning to Discover Novel Interactions in
Biological Swarms [3.441021278275805]
We propose training deep network models to predict system-level states directly from generic graphical features from the entire view.
Because the resulting predictive models are not based on human-understood predictors, we use explanatory modules.
This represents an example of augmented intelligence in behavioral ecology -- knowledge co-creation in a human-AI team.
arXiv Detail & Related papers (2021-08-20T22:50:41Z) - Adversarial Intrinsic Motivation for Reinforcement Learning [60.322878138199364]
We investigate whether the Wasserstein-1 distance between a policy's state visitation distribution and a target distribution can be utilized effectively for reinforcement learning tasks.
Our approach, termed Adversarial Intrinsic Motivation (AIM), estimates this Wasserstein-1 distance through its dual objective and uses it to compute a supplemental reward function.
arXiv Detail & Related papers (2021-05-27T17:51:34Z) - Maximizing Information Gain in Partially Observable Environments via
Prediction Reward [64.24528565312463]
This paper tackles the challenge of using belief-based rewards for a deep RL agent.
We derive the exact error between negative entropy and the expected prediction reward.
This insight provides theoretical motivation for several fields using prediction rewards.
arXiv Detail & Related papers (2020-05-11T08:13:49Z) - Understanding Global Feature Contributions With Additive Importance
Measures [14.50261153230204]
We explore the perspective of defining feature importance through the predictive power associated with each feature.
We introduce two notions of predictive power (model-based and universal) and formalize this approach with a framework of additive importance measures.
We then propose SAGE, a model-agnostic method that quantifies predictive power while accounting for feature interactions.
arXiv Detail & Related papers (2020-04-01T19:17:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.