An intelligent tutor for planning in large partially observable environments
- URL: http://arxiv.org/abs/2302.02785v2
- Date: Thu, 6 Jun 2024 13:29:08 GMT
- Title: An intelligent tutor for planning in large partially observable environments
- Authors: Lovis Heindrich, Saksham Consul, Falk Lieder,
- Abstract summary: We develop and evaluate the first intelligent tutor for planning in partially observable environments.
Compared to previous intelligent tutors for teaching planning strategies, this novel intelligent tutor combines two innovations.
A preregistered experiment with 330 participants demonstrated that the new intelligent tutor is highly effective at improving people's ability to make good decisions in partially observable environments.
- Score: 0.8739101659113157
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: AI can not only outperform people in many planning tasks, but it can also teach them how to plan better. A recent and promising approach to improving human decision-making is to create intelligent tutors that utilize AI to discover and teach optimal planning strategies automatically. Prior work has shown that this approach can improve planning in artificial, fully observable planning tasks. Unlike these artificial tasks, the world is only partially observable. To bridge this gap, we developed and evaluated the first intelligent tutor for planning in partially observable environments. Compared to previous intelligent tutors for teaching planning strategies, this novel intelligent tutor combines two innovations: 1) a new metareasoning algorithm for discovering optimal planning strategies for large, partially observable environments, and 2) scaffolding the learning processing by having the learner choose from an increasing larger set of planning operations in increasingly larger planning problems. We found that our new strategy discovery algorithm is superior to the state-of-the-art. A preregistered experiment with 330 participants demonstrated that the new intelligent tutor is highly effective at improving people's ability to make good decisions in partially observable environments. This suggests our human-centered tutoring approach can successfully boost human planning in complex, partially observable sequential decision problems, a promising step towards using AI-powered intelligent tutors to improve human planning in the real world.
Related papers
- SPIRE: Synergistic Planning, Imitation, and Reinforcement Learning for Long-Horizon Manipulation [58.14969377419633]
We propose spire, a system that decomposes tasks into smaller learning subproblems and second combines imitation and reinforcement learning to maximize their strengths.
We find that spire outperforms prior approaches that integrate imitation learning, reinforcement learning, and planning by 35% to 50% in average task performance.
arXiv Detail & Related papers (2024-10-23T17:42:07Z) - Leveraging automatic strategy discovery to teach people how to select better projects [0.9821874476902969]
The decisions of individuals and organizations are often suboptimal because normative decision strategies are too demanding in the real world.
Recent work suggests that some errors can be prevented by leveraging artificial intelligence to discover and teach prescriptive decision strategies.
This article is the first to extend this approach to a real-world decision problem, namely project selection.
arXiv Detail & Related papers (2024-06-06T13:51:44Z) - Intelligent Execution through Plan Analysis [11.771743106780102]
Planning makes assumptions about the world.
When executing plans, the assumptions are usually not met.
Instead, we focus on the positive impact, or opportunities to find better plans.
Experiments in several paradigmatic robotic tasks show how the approach outperforms standard replanning strategies.
arXiv Detail & Related papers (2024-03-18T18:23:36Z) - Towards Goal-oriented Intelligent Tutoring Systems in Online Education [69.06930979754627]
We propose a new task, named Goal-oriented Intelligent Tutoring Systems (GITS)
GITS aims to enable the student's mastery of a designated concept by strategically planning a customized sequence of exercises and assessment.
We propose a novel graph-based reinforcement learning framework, named Planning-Assessment-Interaction (PAI)
arXiv Detail & Related papers (2023-12-03T12:37:16Z) - Thinker: Learning to Plan and Act [18.425843346728648]
The Thinker algorithm wraps the environment with a world model and introduces new actions designed for interacting with the world model.
We demonstrate the algorithm's effectiveness through experimental results in the game of Sokoban and the Atari 2600 benchmark.
arXiv Detail & Related papers (2023-07-27T16:40:14Z) - Hierarchical Imitation Learning with Vector Quantized Models [77.67190661002691]
We propose to use reinforcement learning to identify subgoals in expert trajectories.
We build a vector-quantized generative model for the identified subgoals to perform subgoal-level planning.
In experiments, the algorithm excels at solving complex, long-horizon decision-making problems outperforming state-of-the-art.
arXiv Detail & Related papers (2023-01-30T15:04:39Z) - Achieving mouse-level strategic evasion performance using real-time
computational planning [59.60094442546867]
Planning is an extraordinary ability in which the brain imagines and then enacts evaluated possible futures.
We develop a more efficient biologically-inspired planning algorithm, TLPPO, based on work on how the ecology of an animal governs the value of spatial planning.
We compare the performance of a real-time agent using TLPPO against the performance of live mice, all tasked with evading a robot predator.
arXiv Detail & Related papers (2022-11-04T18:34:36Z) - Have I done enough planning or should I plan more? [0.7734726150561086]
We show that people acquire this ability through learning and reverse-engineer the underlying learning mechanisms.
We find that people quickly adapt how much planning they perform to the cost and benefit of planning.
Our results suggest that the metacognitive ability to adjust the amount of planning might be learned through a policy-gradient mechanism.
arXiv Detail & Related papers (2022-01-03T17:11:07Z) - Procedure Planning in Instructional Videosvia Contextual Modeling and
Model-based Policy Learning [114.1830997893756]
This work focuses on learning a model to plan goal-directed actions in real-life videos.
We propose novel algorithms to model human behaviors through Bayesian Inference and model-based Imitation Learning.
arXiv Detail & Related papers (2021-10-05T01:06:53Z) - Improving Human Decision-Making by Discovering Efficient Strategies for
Hierarchical Planning [0.6882042556551609]
People need efficient planning strategies because their computational resources are limited.
Our ability to compute those strategies used to be limited to very small and very simple planning tasks.
We introduce a cognitively-inspired reinforcement learning method that can overcome this limitation.
arXiv Detail & Related papers (2021-01-31T19:46:00Z) - Planning to Explore via Self-Supervised World Models [120.31359262226758]
Plan2Explore is a self-supervised reinforcement learning agent.
We present a new approach to self-supervised exploration and fast adaptation to new tasks.
Without any training supervision or task-specific interaction, Plan2Explore outperforms prior self-supervised exploration methods.
arXiv Detail & Related papers (2020-05-12T17:59:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.