Memory-Augmented Theory of Mind Network
- URL: http://arxiv.org/abs/2301.06926v1
- Date: Tue, 17 Jan 2023 14:48:58 GMT
- Title: Memory-Augmented Theory of Mind Network
- Authors: Dung Nguyen, Phuoc Nguyen, Hung Le, Kien Do, Svetha Venkatesh, Truyen
Tran
- Abstract summary: Social reasoning requires the capacity of theory of mind (ToM) to contextualise and attribute mental states to others.
Recent machine learning approaches to ToM have demonstrated that we can train the observer to read the past and present behaviours of other agents.
We tackle the challenges by equipping the observer with novel neural memory mechanisms to encode, and hierarchical attention to selectively retrieve information about others.
This results in ToMMY, a theory of mind model that learns to reason while making little assumptions about the underlying mental processes.
- Score: 59.9781556714202
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Social reasoning necessitates the capacity of theory of mind (ToM), the
ability to contextualise and attribute mental states to others without having
access to their internal cognitive structure. Recent machine learning
approaches to ToM have demonstrated that we can train the observer to read the
past and present behaviours of other agents and infer their beliefs (including
false beliefs about things that no longer exist), goals, intentions and future
actions. The challenges arise when the behavioural space is complex, demanding
skilful space navigation for rapidly changing contexts for an extended period.
We tackle the challenges by equipping the observer with novel neural memory
mechanisms to encode, and hierarchical attention to selectively retrieve
information about others. The memories allow rapid, selective querying of
distal related past behaviours of others to deliberatively reason about their
current mental state, beliefs and future behaviours. This results in ToMMY, a
theory of mind model that learns to reason while making little assumptions
about the underlying mental processes. We also construct a new suite of
experiments to demonstrate that memories facilitate the learning process and
achieve better theory of mind performance, especially for high-demand
false-belief tasks that require inferring through multiple steps of changes.
Related papers
- Learning mental states estimation through self-observation: a developmental synergy between intentions and beliefs representations in a deep-learning model of Theory of Mind [0.35154948148425685]
Theory of Mind (ToM) is the ability to attribute beliefs, intentions, or mental states to others.
We show a developmental synergy between learning to predict low-level mental states and attributing high-level ones.
We propose that our computational approach can inform the understanding of human social cognitive development.
arXiv Detail & Related papers (2024-07-25T13:15:25Z) - Think Twice: Perspective-Taking Improves Large Language Models'
Theory-of-Mind Capabilities [63.90227161974381]
SimToM is a novel prompting framework inspired by Simulation Theory's notion of perspective-taking.
Our approach, which requires no additional training and minimal prompt-tuning, shows substantial improvement over existing methods.
arXiv Detail & Related papers (2023-11-16T22:49:27Z) - Neural Theory-of-Mind? On the Limits of Social Intelligence in Large LMs [77.88043871260466]
We show that one of today's largest language models lacks this kind of social intelligence out-of-the box.
We conclude that person-centric NLP approaches might be more effective towards neural Theory of Mind.
arXiv Detail & Related papers (2022-10-24T14:58:58Z) - Learning Theory of Mind via Dynamic Traits Attribution [59.9781556714202]
We propose a new neural ToM architecture that learns to generate a latent trait vector of an actor from the past trajectories.
This trait vector then multiplicatively modulates the prediction mechanism via a fast weights' scheme in the prediction neural network.
We empirically show that the fast weights provide a good inductive bias to model the character traits of agents and hence improves mindreading ability.
arXiv Detail & Related papers (2022-04-17T11:21:18Z) - Mind the gap: Challenges of deep learning approaches to Theory of Mind [0.0]
Theory of Mind is an essential ability of humans to infer the mental states of others.
Here we provide a coherent summary of the potential, current progress, and problems of deep learning approaches to Theory of Mind.
arXiv Detail & Related papers (2022-03-30T15:48:05Z) - AGENT: A Benchmark for Core Psychological Reasoning [60.35621718321559]
Intuitive psychology is the ability to reason about hidden mental variables that drive observable actions.
Despite recent interest in machine agents that reason about other agents, it is not clear if such agents learn or hold the core psychology principles that drive human reasoning.
We present a benchmark consisting of procedurally generated 3D animations, AGENT, structured around four scenarios.
arXiv Detail & Related papers (2021-02-24T14:58:23Z) - Episodic memory governs choices: An RNN-based reinforcement learning
model for decision-making task [24.96447960548042]
We develop an RNN-based Actor-Critic framework to solve two tasks analogous to the monkeys' decision-making tasks.
We try to explore an open question in neuroscience: which episodic memory in the hippocampus should be selected to ultimately govern future decisions.
arXiv Detail & Related papers (2021-01-24T04:33:07Z) - Towards a Neural Model for Serial Order in Frontal Cortex: a Brain
Theory from Memory Development to Higher-Level Cognition [53.816853325427424]
We propose that the immature prefrontal cortex (PFC) use its primary functionality of detecting hierarchical patterns in temporal signals.
Our hypothesis is that the PFC detects the hierarchical structure in temporal sequences in the form of ordinal patterns and use them to index information hierarchically in different parts of the brain.
By doing so, it gives the tools to the language-ready brain for manipulating abstract knowledge and planning temporally ordered information.
arXiv Detail & Related papers (2020-05-22T14:29:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.