Parallelized Planning-Acting for Efficient LLM-based Multi-Agent Systems
- URL: http://arxiv.org/abs/2503.03505v1
- Date: Wed, 05 Mar 2025 13:53:10 GMT
- Title: Parallelized Planning-Acting for Efficient LLM-based Multi-Agent Systems
- Authors: Yaoru Li, Shunyu Liu, Tongya Zheng, Mingli Song,
- Abstract summary: We propose a novel parallelized planning-acting framework for Multi-Agent Systems.<n>The proposed framework features a dual-thread architecture with interruptible execution to enable concurrent planning and acting.
- Score: 31.894636711684523
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advancements in Large Language Model(LLM)-based Multi-Agent Systems(MAS) have demonstrated remarkable potential for tackling complex decision-making tasks. However, existing frameworks inevitably rely on serialized execution paradigms, where agents must complete sequential LLM planning before taking action. This fundamental constraint severely limits real-time responsiveness and adaptation, which is crucial in dynamic environments with ever-changing scenarios. In this paper, we propose a novel parallelized planning-acting framework for LLM-based MAS, featuring a dual-thread architecture with interruptible execution to enable concurrent planning and acting. Specifically, our framework comprises two core threads:(1) a planning thread driven by a centralized memory system, maintaining synchronization of environmental states and agent communication to support dynamic decision-making; and (2) an acting thread equipped with a comprehensive skill library, enabling automated task execution through recursive decomposition. Extensive experiments on challenging Minecraft demonstrate the effectiveness of the proposed framework.
Related papers
- SagaLLM: Context Management, Validation, and Transaction Guarantees for Multi-Agent LLM Planning [2.1331883629523634]
SagaLLM is a structured multi-agent framework that addresses four fundamental limitations in current LLM approaches.
By implementing specialized context management agents and validation protocols, SagaLLM preserves critical constraints and state information throughout complex planning processes.
arXiv Detail & Related papers (2025-03-15T01:43:03Z) - DynTaskMAS: A Dynamic Task Graph-driven Framework for Asynchronous and Parallel LLM-based Multi-Agent Systems [2.6353853440763113]
This paper introduces DynTaskMAS, a novel framework that orchestrates asynchronous and parallel operations in Multi-Agent Systems.
The framework features four key innovations: (1) a Dynamic Task Graph Generator that decomposes complex tasks while maintaining logical dependencies, (2) an Asynchronous Parallel Execution Engine that optimize resource utilization through efficient task scheduling, and (3) a Semantic-Aware Context Management System that enables efficient information sharing among agents.
arXiv Detail & Related papers (2025-03-10T06:16:10Z) - A Temporal Planning Framework for Multi-Agent Systems via LLM-Aided Knowledge Base Management [5.548477348501636]
This paper presents a novel framework, called PLANTOR, that integrates Large Language Models (LLMs) with Prolog-based knowledge management and planning for multi-robot tasks.<n>Results demonstrate that LLMs can produce accurate knowledge bases with modest human feedback, while Prolog guarantees formal correctness and explainability.<n>This approach underscores the potential of LLM integration for advanced robotics tasks requiring flexible, scalable, and human-understandable planning.
arXiv Detail & Related papers (2025-02-26T13:51:28Z) - REALM-Bench: A Real-World Planning Benchmark for LLMs and Multi-Agent Systems [2.1331883629523634]
The suite encompasses eleven designed problems that progress from basic to highly complex.<n>Each problem can be scaled along three dimensions: the number of parallel planning threads, the complexity of inter-dependencies, and the frequency of unexpected disruptions.<n>The benchmark aims to drive progress in developing more robust and adaptable AI planning systems for real-world applications.
arXiv Detail & Related papers (2025-02-26T05:24:22Z) - Flow: Modularized Agentic Workflow Automation [53.073598156915615]
Multi-agent frameworks powered by large language models (LLMs) have demonstrated great success in automated planning and task execution.
However, the effective adjustment of agentic during execution has not been well studied.
In this paper, we define an activity-on-vertex (AOV) graph, which allows continuous workflow refinement by agents.
Our proposed multi-agent framework achieves efficient concurrent execution of subtasks, effective goal achievement, and enhanced error tolerance.
arXiv Detail & Related papers (2025-01-14T04:35:37Z) - Ontology-driven Prompt Tuning for LLM-based Task and Motion Planning [0.20940572815908076]
Task and Motion Planning (TAMP) approaches combine high-level symbolic plan with low-level motion planning.<n>LLMs are transforming task planning by offering natural language as an intuitive and flexible way to describe tasks.<n>This work proposes a novel prompt-tuning framework that employs knowledge-based reasoning to refine and expand user prompts.
arXiv Detail & Related papers (2024-12-10T13:18:45Z) - Interactive and Expressive Code-Augmented Planning with Large Language Models [62.799579304821826]
Large Language Models (LLMs) demonstrate strong abilities in common-sense reasoning and interactive decision-making.
Recent techniques have sought to structure LLM outputs using control flow and other code-adjacent techniques to improve planning performance.
We propose REPL-Plan, an LLM planning approach that is fully code-expressive and dynamic.
arXiv Detail & Related papers (2024-11-21T04:23:17Z) - Unlocking Reasoning Potential in Large Langauge Models by Scaling Code-form Planning [94.76546523689113]
We introduce CodePlan, a framework that generates and follows textcode-form plans -- pseudocode that outlines high-level, structured reasoning processes.
CodePlan effectively captures the rich semantics and control flows inherent to sophisticated reasoning tasks.
It achieves a 25.1% relative improvement compared with directly generating responses.
arXiv Detail & Related papers (2024-09-19T04:13:58Z) - CoAct: A Global-Local Hierarchy for Autonomous Agent Collaboration [87.51781348070914]
Existing LLMs exhibit remarkable performance on various NLP tasks, but still struggle with complex real-world tasks.
We propose the CoAct framework, which transfers the hierarchical planning and collaboration patterns in human society to LLM systems.
arXiv Detail & Related papers (2024-06-19T09:23:53Z) - LLM3:Large Language Model-based Task and Motion Planning with Motion Failure Reasoning [78.2390460278551]
Conventional Task and Motion Planning (TAMP) approaches rely on manually crafted interfaces connecting symbolic task planning with continuous motion generation.
Here, we present LLM3, a novel Large Language Model (LLM)-based TAMP framework featuring a domain-independent interface.
Specifically, we leverage the powerful reasoning and planning capabilities of pre-trained LLMs to propose symbolic action sequences and select continuous action parameters for motion planning.
arXiv Detail & Related papers (2024-03-18T08:03:47Z) - LLMind: Orchestrating AI and IoT with LLM for Complex Task Execution [18.816077341295628]
We present LLMind, a task-oriented AI framework that enables effective collaboration among IoT devices.
Inspired by the functional specialization theory of the brain, our framework integrates an LLM with domain-specific AI modules.
Complex tasks, which may involve collaborations of multiple domain-specific AI modules and IoT devices, are executed through a control script.
arXiv Detail & Related papers (2023-12-14T14:57:58Z) - ADaPT: As-Needed Decomposition and Planning with Language Models [131.063805299796]
We introduce As-Needed Decomposition and Planning for complex Tasks (ADaPT)
ADaPT explicitly plans and decomposes complex sub-tasks as-needed, when the Large Language Models is unable to execute them.
Our results demonstrate that ADaPT substantially outperforms established strong baselines.
arXiv Detail & Related papers (2023-11-08T17:59:15Z) - Improving Planning with Large Language Models: A Modular Agentic Architecture [7.63815864256878]
Large language models (LLMs) often struggle with tasks that require multi-step reasoning or goal-directed planning.
We propose an agentic architecture, the Modular Agentic Planner (MAP), in which planning is accomplished via the recurrent interaction of specialized modules.
We find that MAP yields significant improvements over both standard LLM methods.
arXiv Detail & Related papers (2023-09-30T00:10:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.