The Tyranny of Possibilities in the Design of Task-Oriented LLM Systems:
A Scoping Survey
- URL: http://arxiv.org/abs/2312.17601v1
- Date: Fri, 29 Dec 2023 13:35:20 GMT
- Title: The Tyranny of Possibilities in the Design of Task-Oriented LLM Systems:
A Scoping Survey
- Authors: Dhruv Dhamani and Mary Lou Maher
- Abstract summary: The paper begins by defining a minimal task-oriented LLM system and exploring the design space of such systems.
We discuss a pattern in our results and formulate them into three conjectures.
In all, the scoping survey presents seven conjectures that can help guide future research efforts.
- Score: 1.0489539392650928
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This scoping survey focuses on our current understanding of the design space
for task-oriented LLM systems and elaborates on definitions and relationships
among the available design parameters. The paper begins by defining a minimal
task-oriented LLM system and exploring the design space of such systems through
a thought experiment contemplating the performance of diverse LLM system
configurations (involving single LLMs, single LLM-based agents, and multiple
LLM-based agent systems) on a complex software development task and
hypothesizes the results. We discuss a pattern in our results and formulate
them into three conjectures. While these conjectures may be partly based on
faulty assumptions, they provide a starting point for future research. The
paper then surveys a select few design parameters: covering and organizing
research in LLM augmentation, prompting techniques, and uncertainty estimation,
and discussing their significance. The paper notes the lack of focus on
computational and energy efficiency in evaluating research in these areas. Our
survey findings provide a basis for developing the concept of linear and
non-linear contexts, which we define and use to enable an agent-centric
projection of prompting techniques providing a lens through which prompting
techniques can be viewed as multi-agent systems. The paper discusses the
implications of this lens, for the cross-pollination of research between LLM
prompting and LLM-based multi-agent systems; and also, for the generation of
synthetic training data based on existing prompting techniques in research. In
all, the scoping survey presents seven conjectures that can help guide future
research efforts.
Related papers
- Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search [57.28671084993782]
Large language models (LLMs) have demonstrated remarkable reasoning capabilities across diverse domains.
Recent studies have shown that increasing test-time computation enhances LLMs' reasoning capabilities.
We propose a two-stage training paradigm: 1) a small-scale format tuning stage to internalize the COAT reasoning format and 2) a large-scale self-improvement stage leveraging reinforcement learning.
arXiv Detail & Related papers (2025-02-04T17:26:58Z) - A Framework for Using LLMs for Repository Mining Studies in Empirical Software Engineering [12.504438766461027]
Large Language Models (LLMs) have transformed Software Engineering (SE) by providing innovative methods for analyzing software repositories.
Our research packages a framework, coined Prompt Refinement and Insights for Mining Empirical Software repositories (PRIMES)
Our findings indicate that standardizing prompt engineering and using PRIMES can enhance the reliability and accuracy of studies utilizing LLMs.
arXiv Detail & Related papers (2024-11-15T06:08:57Z) - Beyond Binary: Towards Fine-Grained LLM-Generated Text Detection via Role Recognition and Involvement Measurement [51.601916604301685]
Large language models (LLMs) generate content that can undermine trust in online discourse.
Current methods often focus on binary classification, failing to address the complexities of real-world scenarios like human-LLM collaboration.
To move beyond binary classification and address these challenges, we propose a new paradigm for detecting LLM-generated content.
arXiv Detail & Related papers (2024-10-18T08:14:10Z) - EVOLvE: Evaluating and Optimizing LLMs For Exploration [76.66831821738927]
Large language models (LLMs) remain under-studied in scenarios requiring optimal decision-making under uncertainty.
We measure LLMs' (in)ability to make optimal decisions in bandits, a state-less reinforcement learning setting relevant to many applications.
Motivated by the existence of optimal exploration algorithms, we propose efficient ways to integrate this algorithmic knowledge into LLMs.
arXiv Detail & Related papers (2024-10-08T17:54:03Z) - Efficient Prompting for LLM-based Generative Internet of Things [88.84327500311464]
Large language models (LLMs) have demonstrated remarkable capacities on various tasks, and integrating the capacities of LLMs into the Internet of Things (IoT) applications has drawn much research attention recently.
Due to security concerns, many institutions avoid accessing state-of-the-art commercial LLM services, requiring the deployment and utilization of open-source LLMs in a local network setting.
We propose a LLM-based Generative IoT (GIoT) system deployed in the local network setting in this study.
arXiv Detail & Related papers (2024-06-14T19:24:00Z) - Towards Modeling Learner Performance with Large Language Models [7.002923425715133]
This paper investigates whether the pattern recognition and sequence modeling capabilities of LLMs can be extended to the domain of knowledge tracing.
We compare two approaches to using LLMs for this task, zero-shot prompting and model fine-tuning, with existing, non-LLM approaches to knowledge tracing.
While LLM-based approaches do not achieve state-of-the-art performance, fine-tuned LLMs surpass the performance of naive baseline models and perform on par with standard Bayesian Knowledge Tracing approaches.
arXiv Detail & Related papers (2024-02-29T14:06:34Z) - LLM Inference Unveiled: Survey and Roofline Model Insights [62.92811060490876]
Large Language Model (LLM) inference is rapidly evolving, presenting a unique blend of opportunities and challenges.
Our survey stands out from traditional literature reviews by not only summarizing the current state of research but also by introducing a framework based on roofline model.
This framework identifies the bottlenecks when deploying LLMs on hardware devices and provides a clear understanding of practical problems.
arXiv Detail & Related papers (2024-02-26T07:33:05Z) - Understanding the planning of LLM agents: A survey [98.82513390811148]
This survey provides the first systematic view of LLM-based agents planning, covering recent works aiming to improve planning ability.
Comprehensive analyses are conducted for each direction, and further challenges in the field of research are discussed.
arXiv Detail & Related papers (2024-02-05T04:25:24Z) - Assessing Logical Puzzle Solving in Large Language Models: Insights from a Minesweeper Case Study [10.95835611110119]
We introduce a novel task -- Minesweeper -- designed in a format unfamiliar to Large Language Models (LLMs)
This task challenges LLMs to identify the locations of mines based on numerical clues provided by adjacent opened cells.
Our experiments, including trials with the advanced GPT-4 model, indicate that while LLMs possess the foundational abilities required for this task, they struggle to integrate these into a coherent, multi-step logical reasoning process needed to solve Minesweeper.
arXiv Detail & Related papers (2023-11-13T15:11:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.