Investigate-Consolidate-Exploit: A General Strategy for Inter-Task Agent
Self-Evolution
- URL: http://arxiv.org/abs/2401.13996v1
- Date: Thu, 25 Jan 2024 07:47:49 GMT
- Title: Investigate-Consolidate-Exploit: A General Strategy for Inter-Task Agent
Self-Evolution
- Authors: Cheng Qian, Shihao Liang, Yujia Qin, Yining Ye, Xin Cong, Yankai Lin,
Yesai Wu, Zhiyuan Liu, Maosong Sun
- Abstract summary: Investigate-Consolidate-Exploit (ICE) is a novel strategy for enhancing the adaptability and flexibility of AI agents.
ICE promotes the transfer of knowledge between tasks for genuine self-evolution.
Our experiments on the XAgent framework demonstrate ICE's effectiveness, reducing API calls by as much as 80%.
- Score: 92.84441068115517
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces Investigate-Consolidate-Exploit (ICE), a novel strategy
for enhancing the adaptability and flexibility of AI agents through inter-task
self-evolution. Unlike existing methods focused on intra-task learning, ICE
promotes the transfer of knowledge between tasks for genuine self-evolution,
similar to human experience learning. The strategy dynamically investigates
planning and execution trajectories, consolidates them into simplified
workflows and pipelines, and exploits them for improved task execution. Our
experiments on the XAgent framework demonstrate ICE's effectiveness, reducing
API calls by as much as 80% and significantly decreasing the demand for the
model's capability. Specifically, when combined with GPT-3.5, ICE's performance
matches that of raw GPT-4 across various agent tasks. We argue that this
self-evolution approach represents a paradigm shift in agent design,
contributing to a more robust AI community and ecosystem, and moving a step
closer to full autonomy.
Related papers
- Intelligent Mobile AI-Generated Content Services via Interactive Prompt Engineering and Dynamic Service Provisioning [55.641299901038316]
AI-generated content can organize collaborative Mobile AIGC Service Providers (MASPs) at network edges to provide ubiquitous and customized content for resource-constrained users.
Such a paradigm faces two significant challenges: 1) raw prompts often lead to poor generation quality due to users' lack of experience with specific AIGC models, and 2) static service provisioning fails to efficiently utilize computational and communication resources.
We develop an interactive prompt engineering mechanism that leverages a Large Language Model (LLM) to generate customized prompt corpora and employs Inverse Reinforcement Learning (IRL) for policy imitation.
arXiv Detail & Related papers (2025-02-17T03:05:20Z) - From Novice to Expert: LLM Agent Policy Optimization via Step-wise Reinforcement Learning [62.54484062185869]
We introduce StepAgent, which utilizes step-wise reward to optimize the agent's reinforcement learning process.
We propose implicit-reward and inverse reinforcement learning techniques to facilitate agent reflection and policy adjustment.
arXiv Detail & Related papers (2024-11-06T10:35:11Z) - Meta-Learning Integration in Hierarchical Reinforcement Learning for Advanced Task Complexity [0.0]
Hierarchical Reinforcement Learning (HRL) effectively tackles complex tasks by decomposing them into structured policies.
We integrate meta-learning into HRL to enhance the agent's ability to learn and adapt hierarchical policies swiftly.
arXiv Detail & Related papers (2024-10-10T13:47:37Z) - Gödel Agent: A Self-Referential Agent Framework for Recursive Self-Improvement [117.94654815220404]
G"odel Agent is a self-evolving framework inspired by the G"odel machine.
G"odel Agent can achieve continuous self-improvement, surpassing manually crafted agents in performance, efficiency, and generalizability.
arXiv Detail & Related papers (2024-10-06T10:49:40Z) - Textualized Agent-Style Reasoning for Complex Tasks by Multiple Round LLM Generation [49.27250832754313]
We present AgentCOT, a llm-based autonomous agent framework.
At each step, AgentCOT selects an action and executes it to yield an intermediate result with supporting evidence.
We introduce two new strategies to enhance the performance of AgentCOT.
arXiv Detail & Related papers (2024-09-19T02:20:06Z) - Instigating Cooperation among LLM Agents Using Adaptive Information Modulation [0.6749750044497731]
This paper introduces a novel framework combining LLM agents as proxies for human strategic behavior with reinforcement learning.
We show that pro-social promoting RL agent modulates information access across agents in a network, optimizing social welfare and promoting pro-social behavior.
This framework offers significant insights into AI-mediated social dynamics, contributing to the deployment of AI in real-world team settings.
arXiv Detail & Related papers (2024-09-16T15:15:51Z) - Tradeoffs When Considering Deep Reinforcement Learning for Contingency Management in Advanced Air Mobility [0.0]
Air transportation is undergoing a rapid evolution globally with the introduction of Advanced Air Mobility (AAM)
Increased levels of automation are likely necessary to achieve operational safety and efficiency goals.
This paper explores the use of Deep Reinforcement Learning (DRL) which has shown promising performance in complex and high-dimensional environments.
arXiv Detail & Related papers (2024-06-28T19:09:55Z) - Learning to Transfer for Evolutionary Multitasking [30.01466615418299]
Evolutionary multitasking (EMT) is an emerging approach for solving multitask optimization problems (MTOPs)
Current approaches in implicit EMT face challenges in adaptability, due to the use of a limited number of evolution operators.
We propose a novel Learning to Transfer (L2T) framework to automatically discover efficient KT policies for the MTOPs.
arXiv Detail & Related papers (2024-06-20T14:31:24Z) - Decentralized and Lifelong-Adaptive Multi-Agent Collaborative Learning [57.652899266553035]
Decentralized and lifelong-adaptive multi-agent collaborative learning aims to enhance collaboration among multiple agents without a central server.
We propose DeLAMA, a decentralized multi-agent lifelong collaborative learning algorithm with dynamic collaboration graphs.
arXiv Detail & Related papers (2024-03-11T09:21:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.