IRRGN: An Implicit Relational Reasoning Graph Network for Multi-turn
Response Selection
- URL: http://arxiv.org/abs/2212.00482v2
- Date: Mon, 23 Oct 2023 12:16:52 GMT
- Title: IRRGN: An Implicit Relational Reasoning Graph Network for Multi-turn
Response Selection
- Authors: Jingcheng Deng, Hengwei Dai, Xuewei Guo, Yuanchen Ju and Wei Peng
- Abstract summary: Implicit Reasoning to Graph Network aims to implicitly extract between utterances, as well as utterances and options.
Model surpasses human performance for the first time on the MuTual dataset.
- Score: 4.471148909362883
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The task of response selection in multi-turn dialogue is to find the best
option from all candidates. In order to improve the reasoning ability of the
model, previous studies pay more attention to using explicit algorithms to
model the dependencies between utterances, which are deterministic, limited and
inflexible. In addition, few studies consider differences between the options
before and after reasoning. In this paper, we propose an Implicit Relational
Reasoning Graph Network to address these issues, which consists of the
Utterance Relational Reasoner (URR) and the Option Dual Comparator (ODC). URR
aims to implicitly extract dependencies between utterances, as well as
utterances and options, and make reasoning with relational graph convolutional
networks. ODC focuses on perceiving the difference between the options through
dual comparison, which can eliminate the interference of the noise options.
Experimental results on two multi-turn dialogue reasoning benchmark datasets
MuTual and MuTual+ show that our method significantly improves the baseline of
four pretrained language models and achieves state-of-the-art performance. The
model surpasses human performance for the first time on the MuTual dataset.
Related papers
- Vision-Language Models Can Self-Improve Reasoning via Reflection [20.196406628954303]
Chain-of-thought (CoT) has proven to improve the reasoning capability of large language models (LLMs)
We propose a self-training framework, R3V, which iteratively enhances the model's Vision-language Reasoning by Reflecting on CoT Rationales.
Our approach supports self-reflection on generated solutions, further boosting performance through test-time computation.
arXiv Detail & Related papers (2024-10-30T14:45:00Z) - Coreference-aware Double-channel Attention Network for Multi-party
Dialogue Reading Comprehension [7.353227696624305]
We tackle Multi-party Dialogue Reading (abbr., MDRC)
MDRC stands for an extractive reading comprehension task grounded on a batch of dialogues among multiple interlocutors.
We propose a coreference-aware attention modeling method to strengthen the reasoning ability.
arXiv Detail & Related papers (2023-05-15T05:01:29Z) - Selective Inference for Sparse Multitask Regression with Applications in
Neuroimaging [2.611153304251067]
We propose a framework for selective inference to address a common multi-task problem in neuroimaging.
Our framework offers a new conditional procedure for inference, based on a refinement of the selection event that yields a tractable selection-adjusted likelihood.
We demonstrate through simulations that multi-task learning with selective inference can more accurately recover true signals than single-task methods.
arXiv Detail & Related papers (2022-05-27T20:21:20Z) - Visualizing the Relationship Between Encoded Linguistic Information and
Task Performance [53.223789395577796]
We study the dynamic relationship between the encoded linguistic information and task performance from the viewpoint of Pareto Optimality.
We conduct experiments on two popular NLP tasks, i.e., machine translation and language modeling, and investigate the relationship between several kinds of linguistic information and task performances.
Our empirical findings suggest that some syntactic information is helpful for NLP tasks whereas encoding more syntactic information does not necessarily lead to better performance.
arXiv Detail & Related papers (2022-03-29T19:03:10Z) - Learning MDPs from Features: Predict-Then-Optimize for Sequential
Decision Problems by Reinforcement Learning [52.74071439183113]
We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) solved via reinforcement learning.
Two significant computational challenges arise in applying decision-focused learning to MDPs.
arXiv Detail & Related papers (2021-06-06T23:53:31Z) - Auto-weighted Multi-view Feature Selection with Graph Optimization [90.26124046530319]
We propose a novel unsupervised multi-view feature selection model based on graph learning.
The contributions are threefold: (1) during the feature selection procedure, the consensus similarity graph shared by different views is learned.
Experiments on various datasets demonstrate the superiority of the proposed method compared with the state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T03:25:25Z) - A Graph Reasoning Network for Multi-turn Response Selection via
Customized Pre-training [11.532734330690584]
We propose a graph-reasoning network (GRN) to address the problem.
GRN first conducts pre-training based on ALBERT.
We then fine-tune the model on an integrated network with sequence reasoning and graph reasoning structures.
arXiv Detail & Related papers (2020-12-21T03:38:29Z) - Learning an Effective Context-Response Matching Model with
Self-Supervised Tasks for Retrieval-based Dialogues [88.73739515457116]
We introduce four self-supervised tasks including next session prediction, utterance restoration, incoherence detection and consistency discrimination.
We jointly train the PLM-based response selection model with these auxiliary tasks in a multi-task manner.
Experiment results indicate that the proposed auxiliary self-supervised tasks bring significant improvement for multi-turn response selection.
arXiv Detail & Related papers (2020-09-14T08:44:46Z) - Improving Multi-Turn Response Selection Models with Complementary
Last-Utterance Selection by Instance Weighting [84.9716460244444]
We consider utilizing the underlying correlation in the data resource itself to derive different kinds of supervision signals.
We conduct extensive experiments in two public datasets and obtain significant improvement in both datasets.
arXiv Detail & Related papers (2020-02-18T06:29:01Z) - Joint Contextual Modeling for ASR Correction and Language Understanding [60.230013453699975]
We propose multi-task neural approaches to perform contextual language correction on ASR outputs jointly with language understanding (LU)
We show that the error rates of off the shelf ASR and following LU systems can be reduced significantly by 14% relative with joint models trained using small amounts of in-domain data.
arXiv Detail & Related papers (2020-01-28T22:09:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.