Evaluating Pretrained Transformer Models for Entity Linking in
Task-Oriented Dialog
- URL: http://arxiv.org/abs/2112.08327v1
- Date: Wed, 15 Dec 2021 18:20:12 GMT
- Title: Evaluating Pretrained Transformer Models for Entity Linking in
Task-Oriented Dialog
- Authors: Sai Muralidhar Jayanthi, Varsha Embar, Karthik Raghunathan
- Abstract summary: We evaluate different pretrained transformer models (PTMs) for understanding short phrases of text.
Several of the PTMs produce sub-par results when compared to traditional techniques.
We find that some of their shortcomings can be addressed by using PTMs fine-tuned for text-similarity tasks.
- Score: 1.4524096882720263
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The wide applicability of pretrained transformer models (PTMs) for natural
language tasks is well demonstrated, but their ability to comprehend short
phrases of text is less explored. To this end, we evaluate different PTMs from
the lens of unsupervised Entity Linking in task-oriented dialog across 5
characteristics -- syntactic, semantic, short-forms, numeric and phonetic. Our
results demonstrate that several of the PTMs produce sub-par results when
compared to traditional techniques, albeit competitive to other neural
baselines. We find that some of their shortcomings can be addressed by using
PTMs fine-tuned for text-similarity tasks, which illustrate an improved ability
in comprehending semantic and syntactic correspondences, as well as some
improvements for short-forms, numeric and phonetic variations in entity
mentions. We perform qualitative analysis to understand nuances in their
predictions and discuss scope for further improvements. Code can be found at
https://github.com/murali1996/el_tod
Related papers
- Beyond Coarse-Grained Matching in Video-Text Retrieval [50.799697216533914]
We introduce a new approach for fine-grained evaluation.
Our approach can be applied to existing datasets by automatically generating hard negative test captions.
Experiments on our fine-grained evaluations demonstrate that this approach enhances a model's ability to understand fine-grained differences.
arXiv Detail & Related papers (2024-10-16T09:42:29Z) - Evaluating Semantic Variation in Text-to-Image Synthesis: A Causal Perspective [50.261681681643076]
We propose a novel metric called SemVarEffect and a benchmark named SemVarBench to evaluate the causality between semantic variations in inputs and outputs in text-to-image synthesis.
Our work establishes an effective evaluation framework that advances the T2I synthesis community's exploration of human instruction understanding.
arXiv Detail & Related papers (2024-10-14T08:45:35Z) - Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - Improving Language Models Meaning Understanding and Consistency by
Learning Conceptual Roles from Dictionary [65.268245109828]
Non-human-like behaviour of contemporary pre-trained language models (PLMs) is a leading cause undermining their trustworthiness.
A striking phenomenon is the generation of inconsistent predictions, which produces contradictory results.
We propose a practical approach that alleviates the inconsistent behaviour issue by improving PLM awareness.
arXiv Detail & Related papers (2023-10-24T06:15:15Z) - On Conditional and Compositional Language Model Differentiable Prompting [75.76546041094436]
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks.
We propose a new model, Prompt Production System (PRopS), which learns to transform task instructions or input metadata, into continuous prompts.
arXiv Detail & Related papers (2023-07-04T02:47:42Z) - Teaching Probabilistic Logical Reasoning to Transformers [21.335836561959887]
We evaluate the capability of transformer-based language models in making inferences over uncertain text.
We propose a novel end-to-end fine-tuning approach, Probabilistic Constraint Training.
arXiv Detail & Related papers (2023-05-22T16:08:20Z) - Stabilized In-Context Learning with Pre-trained Language Models for Few
Shot Dialogue State Tracking [57.92608483099916]
Large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks.
For more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial.
We introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query.
arXiv Detail & Related papers (2023-02-12T15:05:10Z) - To Augment or Not to Augment? A Comparative Study on Text Augmentation
Techniques for Low-Resource NLP [0.0]
We investigate three categories of text augmentation methodologies which perform changes on the syntax.
We compare them on part-of-speech tagging, dependency parsing and semantic role labeling for a diverse set of language families.
Our results suggest that the augmentation techniques can further improve over strong baselines based on mBERT.
arXiv Detail & Related papers (2021-11-18T10:52:48Z) - Training Bi-Encoders for Word Sense Disambiguation [4.149972584899897]
State-of-the-art approaches in Word Sense Disambiguation leverage lexical information along with pre-trained embeddings from these models to achieve results comparable to human inter-annotator agreement on standard evaluation benchmarks.
We further the state of the art in Word Sense Disambiguation through our multi-stage pre-training and fine-tuning pipeline.
arXiv Detail & Related papers (2021-05-21T06:06:03Z) - Analysis and Evaluation of Language Models for Word Sense Disambiguation [18.001457030065712]
Transformer-based language models have taken many fields in NLP by storm.
BERT can accurately capture high-level sense distinctions, even when a limited number of examples is available for each word sense.
BERT and its derivatives dominate most of the existing evaluation benchmarks.
arXiv Detail & Related papers (2020-08-26T15:07:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.