A Review on Machine Theory of Mind
- URL: http://arxiv.org/abs/2303.11594v1
- Date: Tue, 21 Mar 2023 04:58:47 GMT
- Title: A Review on Machine Theory of Mind
- Authors: Yuanyuan Mao, Shuang Liu, Pengshuai Zhao, Qin Ni, Xin Lin and Liang He
- Abstract summary: Theory of Mind (ToM) is the ability to attribute mental states to others, the basis of human cognition.
In this paper, we review recent progress in machine ToM on beliefs, desires, and intentions.
- Score: 16.967933605635203
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Theory of Mind (ToM) is the ability to attribute mental states to others, the
basis of human cognition. At present, there has been growing interest in the AI
with cognitive abilities, for example in healthcare and the motoring industry.
Beliefs, desires, and intentions are the early abilities of infants and the
foundation of human cognitive ability, as well as for machine with ToM. In this
paper, we review recent progress in machine ToM on beliefs, desires, and
intentions. And we shall introduce the experiments, datasets and methods of
machine ToM on these three aspects, summarize the development of different
tasks and datasets in recent years, and compare well-behaved models in aspects
of advantages, limitations and applicable conditions, hoping that this study
can guide researchers to quickly keep up with latest trend in this field.
Unlike other domains with a specific task and resolution framework, machine ToM
lacks a unified instruction and a series of standard evaluation tasks, which
make it difficult to formally compare the proposed models. We argue that, one
method to address this difficulty is now to present a standard assessment
criteria and dataset, better a large-scale dataset covered multiple aspects of
ToM.
Related papers
- MFE-ETP: A Comprehensive Evaluation Benchmark for Multi-modal Foundation Models on Embodied Task Planning [50.45558735526665]
We provide an in-depth and comprehensive evaluation of the performance of MFMs on embodied task planning.
We propose a new benchmark, named MFE-ETP, characterized its complex and variable task scenarios.
Using the benchmark and evaluation platform, we evaluated several state-of-the-art MFMs and found that they significantly lag behind human-level performance.
arXiv Detail & Related papers (2024-07-06T11:07:18Z) - NegotiationToM: A Benchmark for Stress-testing Machine Theory of Mind on Negotiation Surrounding [55.38254464415964]
Theory of mind evaluations currently focuses on testing models using machine-generated data or game settings prone to shortcuts and spurious correlations.
We introduce NegotiationToM, a new benchmark designed to stress-test machine ToM in real-world negotiation surrounding covered multi-dimensional mental states.
arXiv Detail & Related papers (2024-04-21T11:51:13Z) - BDIQA: A New Dataset for Video Question Answering to Explore Cognitive
Reasoning through Theory of Mind [21.806678376095576]
Theory of mind (ToM) can make AI more closely resemble human thought processes.
Video question answer (VideoQA) datasets focus on studying causal reasoning within events few of them genuinely incorporating human ToM.
This paper presents BDIQA, the first benchmark to explore the cognitive reasoning capabilities of VideoQA models in the context of ToM.
arXiv Detail & Related papers (2024-02-12T04:34:19Z) - MMToM-QA: Multimodal Theory of Mind Question Answering [80.87550820953236]
Theory of Mind (ToM) is an essential ingredient for developing machines with human-level social intelligence.
Recent machine learning models, particularly large language models, seem to show some aspects of ToM understanding.
Human ToM, on the other hand, is more than video or text understanding.
People can flexibly reason about another person's mind based on conceptual representations extracted from any available data.
arXiv Detail & Related papers (2024-01-16T18:59:24Z) - Think Twice: Perspective-Taking Improves Large Language Models'
Theory-of-Mind Capabilities [63.90227161974381]
SimToM is a novel prompting framework inspired by Simulation Theory's notion of perspective-taking.
Our approach, which requires no additional training and minimal prompt-tuning, shows substantial improvement over existing methods.
arXiv Detail & Related papers (2023-11-16T22:49:27Z) - Towards A Holistic Landscape of Situated Theory of Mind in Large
Language Models [14.491223187047378]
Large Language Models (LLMs) have generated considerable interest and debate regarding their potential emergence of Theory of Mind (ToM)
Several recent inquiries reveal a lack of robust ToM in these models and pose a pressing demand to develop new benchmarks.
We taxonomize machine ToM into 7 mental state categories and delineate existing benchmarks to identify under-explored aspects of ToM.
arXiv Detail & Related papers (2023-10-30T15:12:09Z) - Clever Hans or Neural Theory of Mind? Stress Testing Social Reasoning in
Large Language Models [82.50173296858377]
Many anecdotal examples were used to suggest newer large language models (LLMs) like ChatGPT and GPT-4 exhibit Neural Theory-of-Mind (N-ToM)
We investigate the extent of LLMs' N-ToM through an extensive evaluation on 6 tasks and find that while LLMs exhibit certain N-ToM abilities, this behavior is far from being robust.
arXiv Detail & Related papers (2023-05-24T06:14:31Z) - Machine Psychology [54.287802134327485]
We argue that a fruitful direction for research is engaging large language models in behavioral experiments inspired by psychology.
We highlight theoretical perspectives, experimental paradigms, and computational analysis techniques that this approach brings to the table.
It paves the way for a "machine psychology" for generative artificial intelligence (AI) that goes beyond performance benchmarks.
arXiv Detail & Related papers (2023-03-24T13:24:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.