OMNI-EPIC: Open-endedness via Models of human Notions of Interestingness with Environments Programmed in Code
- URL: http://arxiv.org/abs/2405.15568v2
- Date: Mon, 07 Oct 2024 23:21:20 GMT
- Title: OMNI-EPIC: Open-endedness via Models of human Notions of Interestingness with Environments Programmed in Code
- Authors: Maxence Faldor, Jenny Zhang, Antoine Cully, Jeff Clune,
- Abstract summary: Open-ended and AI-generating algorithms aim to continuously generate and solve increasingly complex tasks indefinitely.
To accomplish this grand vision, learning must occur within a vast array of potential tasks.
We introduce a novel framework, OMNI-EPIC, that augments previous work in Open-endedness.
- Score: 6.067502582087556
- License:
- Abstract: Open-ended and AI-generating algorithms aim to continuously generate and solve increasingly complex tasks indefinitely, offering a promising path toward more general intelligence. To accomplish this grand vision, learning must occur within a vast array of potential tasks. Existing approaches to automatically generating environments are constrained within manually predefined, often narrow distributions of environment, limiting their ability to create any learning environment. To address this limitation, we introduce a novel framework, OMNI-EPIC, that augments previous work in Open-endedness via Models of human Notions of Interestingness (OMNI) with Environments Programmed in Code (EPIC). OMNI-EPIC leverages foundation models to autonomously generate code specifying the next learnable (i.e., not too easy or difficult for the agent's current skill set) and interesting (e.g., worthwhile and novel) tasks. OMNI-EPIC generates both environments (e.g., an obstacle course) and reward functions (e.g., progress through the obstacle course quickly without touching red objects), enabling it, in principle, to create any simulatable learning task. We showcase the explosive creativity of OMNI-EPIC, which continuously innovates to suggest new, interesting learning challenges. We also highlight how OMNI-EPIC can adapt to reinforcement learning agents' learning progress, generating tasks that are of suitable difficulty. Overall, OMNI-EPIC can endlessly create learnable and interesting environments, further propelling the development of self-improving AI systems and AI-Generating Algorithms. Project website with videos: https://dub.sh/omniepic
Related papers
- A Survey on Robotics with Foundation Models: toward Embodied AI [30.999414445286757]
Recent advances in computer vision, natural language processing, and multi-modality learning have shown that the foundation models have superhuman capabilities for specific tasks.
This survey aims to provide a comprehensive and up-to-date overview of foundation models in robotics, focusing on autonomous manipulation and encompassing high-level planning and low-level control.
arXiv Detail & Related papers (2024-02-04T07:55:01Z) - A Survey of Imitation Learning: Algorithms, Recent Developments, and
Challenges [9.288673880680033]
imitation learning (IL) is a process where desired behavior is learned by imitating an expert's behavior.
This paper aims to provide an introduction to IL and an overview of its underlying assumptions and approaches.
It also offers a detailed description of recent advances and emerging areas of research in the field.
arXiv Detail & Related papers (2023-09-05T11:56:07Z) - OpenAGI: When LLM Meets Domain Experts [51.86179657467822]
Human Intelligence (HI) excels at combining basic skills to solve complex tasks.
This capability is vital for Artificial Intelligence (AI) and should be embedded in comprehensive AI Agents.
We introduce OpenAGI, an open-source platform designed for solving multi-step, real-world tasks.
arXiv Detail & Related papers (2023-04-10T03:55:35Z) - Stochastic Coherence Over Attention Trajectory For Continuous Learning
In Video Streams [64.82800502603138]
This paper proposes a novel neural-network-based approach to progressively and autonomously develop pixel-wise representations in a video stream.
The proposed method is based on a human-like attention mechanism that allows the agent to learn by observing what is moving in the attended locations.
Our experiments leverage 3D virtual environments and they show that the proposed agents can learn to distinguish objects just by observing the video stream.
arXiv Detail & Related papers (2022-04-26T09:52:31Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Intrinsically Motivated Goal-Conditioned Reinforcement Learning: a Short
Survey [21.311739361361717]
Developmental approaches argue that learning agents must generate, select and learn to solve their own problems.
Recent years have seen a convergence of developmental approaches and deep reinforcement learning (RL) methods, forming the new domain of developmental machine learning.
This paper proposes a typology of these methods at the intersection of deep RL and developmental approaches, surveys recent approaches and discusses future avenues.
arXiv Detail & Related papers (2020-12-17T18:51:40Z) - AllenAct: A Framework for Embodied AI Research [37.25733386769186]
Embodied AI is in which agents learn to complete tasks through interaction with their environment from egocentric observations.
AllenAct is a modular and flexible learning framework designed with a focus on the unique requirements of Embodied AI research.
arXiv Detail & Related papers (2020-08-28T17:35:22Z) - Learning Compositional Neural Programs for Continuous Control [62.80551956557359]
We propose a novel solution to challenging sparse-reward, continuous control problems.
Our solution, dubbed AlphaNPI-X, involves three separate stages of learning.
We empirically show that AlphaNPI-X can effectively learn to tackle challenging sparse manipulation tasks.
arXiv Detail & Related papers (2020-07-27T08:27:14Z) - Learning with AMIGo: Adversarially Motivated Intrinsic Goals [63.680207855344875]
AMIGo is a goal-generating teacher that proposes Adversarially Motivated Intrinsic Goals.
We show that our method generates a natural curriculum of self-proposed goals which ultimately allows the agent to solve challenging procedurally-generated tasks.
arXiv Detail & Related papers (2020-06-22T10:22:08Z) - AI from concrete to abstract: demystifying artificial intelligence to
the general public [0.0]
This article presents a new methodology, AI from concrete to abstract (AIcon2abs)
The main strategy adopted by is to promote a demystification of artificial intelligence.
The simplicity of the WiSARD weightless artificial neural network model enables easy visualization and understanding of training and classification tasks.
arXiv Detail & Related papers (2020-06-07T01:14:06Z) - Dark, Beyond Deep: A Paradigm Shift to Cognitive AI with Humanlike
Common Sense [142.53911271465344]
We argue that the next generation of AI must embrace "dark" humanlike common sense for solving novel tasks.
We identify functionality, physics, intent, causality, and utility (FPICU) as the five core domains of cognitive AI with humanlike common sense.
arXiv Detail & Related papers (2020-04-20T04:07:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.