GLM-Dialog: Noise-tolerant Pre-training for Knowledge-grounded Dialogue
Generation
- URL: http://arxiv.org/abs/2302.14401v1
- Date: Tue, 28 Feb 2023 08:35:28 GMT
- Title: GLM-Dialog: Noise-tolerant Pre-training for Knowledge-grounded Dialogue
Generation
- Authors: Jing Zhang, Xiaokang Zhang, Daniel Zhang-Li, Jifan Yu, Zijun Yao,
Zeyao Ma, Yiqi Xu, Haohua Wang, Xiaohan Zhang, Nianyi Lin, Sunrui Lu, Juanzi
Li, Jie Tang
- Abstract summary: GLM-Dialog is a large-scale language model (LLM) with 10B parameters capable of knowledge-grounded conversation in Chinese.
We offer our evaluation platform online in an effort to prompt the development of open source models and reliable dialogue evaluation systems.
- Score: 21.91914619107555
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present GLM-Dialog, a large-scale language model (LLM) with 10B parameters
capable of knowledge-grounded conversation in Chinese using a search engine to
access the Internet knowledge. GLM-Dialog offers a series of applicable
techniques for exploiting various external knowledge including both helpful and
noisy knowledge, enabling the creation of robust knowledge-grounded dialogue
LLMs with limited proper datasets. To evaluate the GLM-Dialog more fairly, we
also propose a novel evaluation method to allow humans to converse with
multiple deployed bots simultaneously and compare their performance implicitly
instead of explicitly rating using multidimensional metrics.Comprehensive
evaluations from automatic to human perspective demonstrate the advantages of
GLM-Dialog comparing with existing open source Chinese dialogue models. We
release both the model checkpoint and source code, and also deploy it as a
WeChat application to interact with users. We offer our evaluation platform
online in an effort to prompt the development of open source models and
reliable dialogue evaluation systems. The additional easy-to-use toolkit that
consists of short text entity linking, query generation, and helpful knowledge
classification is also released to enable diverse applications. All the source
code is available on Github.
Related papers
- Self-seeding and Multi-intent Self-instructing LLMs for Generating
Intent-aware Information-Seeking dialogs [18.102066943918473]
Large language models (LLMs) have been shown to be effective in generating synthetic data.
We propose SOLID, which has novel self-seeding and multi-intent self-instructing schemes.
We use SOLID and SOLID-RL to generate more than 300k intent-aware dialogs, surpassing the size of existing datasets.
arXiv Detail & Related papers (2024-02-18T16:20:43Z) - DIALIGHT: Lightweight Multilingual Development and Evaluation of
Task-Oriented Dialogue Systems with Large Language Models [76.79929883963275]
DIALIGHT is a toolkit for developing and evaluating multilingual Task-Oriented Dialogue (ToD) systems.
It features a secure, user-friendly web interface for fine-grained human evaluation at both local utterance level and global dialogue level.
Our evaluations reveal that while PLM fine-tuning leads to higher accuracy and coherence, LLM-based systems excel in producing diverse and likeable responses.
arXiv Detail & Related papers (2024-01-04T11:27:48Z) - DialCLIP: Empowering CLIP as Multi-Modal Dialog Retriever [83.33209603041013]
We propose a parameter-efficient prompt-tuning method named DialCLIP for multi-modal dialog retrieval.
Our approach introduces a multi-modal context generator to learn context features which are distilled into prompts within the pre-trained vision-language model CLIP.
To facilitate various types of retrieval, we also design multiple experts to learn mappings from CLIP outputs to multi-modal representation space.
arXiv Detail & Related papers (2024-01-02T07:40:12Z) - Frugal Prompting for Dialog Models [17.048111072193933]
This study examines different approaches for building dialog systems using large language models (LLMs)
As part of prompt tuning, we experiment with various ways of providing instructions, exemplars, current query and additional context.
The research also analyzes the representations of dialog history that have the optimal usable-information density.
arXiv Detail & Related papers (2023-05-24T09:06:49Z) - Leveraging Large Language Models in Conversational Recommender Systems [9.751217336860924]
A Conversational Recommender System (CRS) offers increased transparency and control to users by enabling them to engage with the system through a real-time multi-turn dialogue.
Large Language Models (LLMs) have exhibited an unprecedented ability to converse naturally and incorporate world knowledge and common-sense reasoning into language understanding.
arXiv Detail & Related papers (2023-05-13T16:40:07Z) - ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented
Instruction Tuning for Digital Human [76.62897301298699]
ChatPLUG is a Chinese open-domain dialogue system for digital human applications that instruction finetunes on a wide range of dialogue tasks in a unified internet-augmented format.
We show that modelname outperforms state-of-the-art Chinese dialogue systems on both automatic and human evaluation.
We deploy modelname to real-world applications such as Smart Speaker and Instant Message applications with fast inference.
arXiv Detail & Related papers (2023-04-16T18:16:35Z) - PK-Chat: Pointer Network Guided Knowledge Driven Generative Dialogue
Model [79.64376762489164]
PK-Chat is a Pointer network guided generative dialogue model, incorporating a unified pretrained language model and a pointer network over knowledge graphs.
The words generated by PK-Chat in the dialogue are derived from the prediction of word lists and the direct prediction of the external knowledge graph knowledge.
Based on the PK-Chat, a dialogue system is built for academic scenarios in the case of geosciences.
arXiv Detail & Related papers (2023-04-02T18:23:13Z) - CGoDial: A Large-Scale Benchmark for Chinese Goal-oriented Dialog
Evaluation [75.60156479374416]
CGoDial is a new challenging and comprehensive Chinese benchmark for Goal-oriented Dialog evaluation.
It contains 96,763 dialog sessions and 574,949 dialog turns totally, covering three datasets with different knowledge sources.
To bridge the gap between academic benchmarks and spoken dialog scenarios, we either collect data from real conversations or add spoken features to existing datasets via crowd-sourcing.
arXiv Detail & Related papers (2022-11-21T16:21:41Z) - Few-Shot Bot: Prompt-Based Learning for Dialogue Systems [58.27337673451943]
Learning to converse using only a few examples is a great challenge in conversational AI.
The current best conversational models are either good chit-chatters (e.g., BlenderBot) or goal-oriented systems (e.g., MinTL)
We propose prompt-based few-shot learning which does not require gradient-based fine-tuning but instead uses a few examples as the only source of learning.
arXiv Detail & Related papers (2021-10-15T14:36:45Z) - Dialog Simulation with Realistic Variations for Training Goal-Oriented
Conversational Systems [14.206866126142002]
Goal-oriented dialog systems enable users to complete specific goals like requesting information about a movie or booking a ticket.
We propose an approach for automatically creating a large corpus of annotated dialogs from a few thoroughly annotated sample dialogs and the dialog schema.
We achieve 18? 50% relative accuracy on a held-out test set compared to a baseline dialog generation approach.
arXiv Detail & Related papers (2020-11-16T19:39:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.