A Communication Theory Perspective on Prompting Engineering Methods for
Large Language Models
- URL: http://arxiv.org/abs/2310.18358v1
- Date: Tue, 24 Oct 2023 03:05:21 GMT
- Title: A Communication Theory Perspective on Prompting Engineering Methods for
Large Language Models
- Authors: Yuanfeng Song, Yuanqin He, Xuefang Zhao, Hanlin Gu, Di Jiang, Haijun
Yang, Lixin Fan, Qiang Yang
- Abstract summary: This article aims to illustrate a novel perspective to review existing prompt engineering (PE) methods, within the well-established communication theory framework.
It aims to facilitate a better/deeper understanding of developing trends of existing PE methods used in four typical tasks.
- Score: 30.57652062704016
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The springing up of Large Language Models (LLMs) has shifted the community
from single-task-orientated natural language processing (NLP) research to a
holistic end-to-end multi-task learning paradigm. Along this line of research
endeavors in the area, LLM-based prompting methods have attracted much
attention, partially due to the technological advantages brought by prompt
engineering (PE) as well as the underlying NLP principles disclosed by various
prompting methods. Traditional supervised learning usually requires training a
model based on labeled data and then making predictions. In contrast, PE
methods directly use the powerful capabilities of existing LLMs (i.e., GPT-3
and GPT-4) via composing appropriate prompts, especially under few-shot or
zero-shot scenarios. Facing the abundance of studies related to the prompting
and the ever-evolving nature of this field, this article aims to (i) illustrate
a novel perspective to review existing PE methods, within the well-established
communication theory framework; (ii) facilitate a better/deeper understanding
of developing trends of existing PE methods used in four typical tasks; (iii)
shed light on promising research directions for future PE methods.
Related papers
- Towards a Unified View of Preference Learning for Large Language Models: A Survey [88.66719962576005]
Large Language Models (LLMs) exhibit remarkably powerful capabilities.
One of the crucial factors to achieve success is aligning the LLM's output with human preferences.
We decompose all the strategies in preference learning into four components: model, data, feedback, and algorithm.
arXiv Detail & Related papers (2024-09-04T15:11:55Z) - Exploring the landscape of large language models: Foundations, techniques, and challenges [8.042562891309414]
The article sheds light on the mechanics of in-context learning and a spectrum of fine-tuning approaches.
It explores how LLMs can be more closely aligned with human preferences through innovative reinforcement learning frameworks.
The ethical dimensions of LLM deployment are discussed, underscoring the need for mindful and responsible application.
arXiv Detail & Related papers (2024-04-18T08:01:20Z) - Simple Techniques for Enhancing Sentence Embeddings in Generative Language Models [3.0566617373924325]
Sentence embedding is a fundamental task within the realm of Natural Language Processing, finding extensive application in search engines, expert systems, and question-and-answer platforms.
With the continuous evolution of large language models such as LLaMA and Mistral, research on sentence embedding has recently achieved notable breakthroughs.
We propose two innovative prompt engineering techniques capable of further enhancing the expressive power of PLMs' raw embeddings.
arXiv Detail & Related papers (2024-04-05T07:07:15Z) - Efficient Prompting Methods for Large Language Models: A Survey [50.171011917404485]
Prompting has become a mainstream paradigm for adapting large language models (LLMs) to specific natural language processing tasks.
This approach brings the additional computational burden of model inference and human effort to guide and control the behavior of LLMs.
We present the basic concepts of prompting, review the advances for efficient prompting, and highlight future research directions.
arXiv Detail & Related papers (2024-04-01T12:19:08Z) - Combatting Human Trafficking in the Cyberspace: A Natural Language
Processing-Based Methodology to Analyze the Language in Online Advertisements [55.2480439325792]
This project tackles the pressing issue of human trafficking in online C2C marketplaces through advanced Natural Language Processing (NLP) techniques.
We introduce a novel methodology for generating pseudo-labeled datasets with minimal supervision, serving as a rich resource for training state-of-the-art NLP models.
A key contribution is the implementation of an interpretability framework using Integrated Gradients, providing explainable insights crucial for law enforcement.
arXiv Detail & Related papers (2023-11-22T02:45:01Z) - Aligning Large Language Models with Human: A Survey [53.6014921995006]
Large Language Models (LLMs) trained on extensive textual corpora have emerged as leading solutions for a broad array of Natural Language Processing (NLP) tasks.
Despite their notable performance, these models are prone to certain limitations such as misunderstanding human instructions, generating potentially biased content, or factually incorrect information.
This survey presents a comprehensive overview of these alignment technologies, including the following aspects.
arXiv Detail & Related papers (2023-07-24T17:44:58Z) - On the Domain Adaptation and Generalization of Pretrained Language
Models: A Survey [15.533482481757353]
We propose a taxonomy of domain adaptation approaches from a machine learning system view.
We discuss and compare those methods and suggest promising future research directions.
arXiv Detail & Related papers (2022-11-06T15:32:00Z) - Meta Learning for Natural Language Processing: A Survey [88.58260839196019]
Deep learning has been the mainstream technique in natural language processing (NLP) area.
Deep learning requires many labeled data and is less generalizable across domains.
Meta-learning is an arising field in machine learning studying approaches to learn better algorithms.
arXiv Detail & Related papers (2022-05-03T13:58:38Z) - Leveraging pre-trained language models for conversational information
seeking from text [2.8425118603312]
In this paper we investigate the usage of in-context learning and pre-trained language representation models to address the problem of information extraction from process description documents.
The results highlight the potential of the approach and the usefulness of the in-context learning customizations.
arXiv Detail & Related papers (2022-03-31T09:00:46Z) - Information Extraction in Low-Resource Scenarios: Survey and Perspective [56.5556523013924]
Information Extraction seeks to derive structured information from unstructured texts.
This paper presents a review of neural approaches to low-resource IE from emphtraditional and emphLLM-based perspectives.
arXiv Detail & Related papers (2022-02-16T13:44:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.