PiC: A Phrase-in-Context Dataset for Phrase Understanding and Semantic
Search
- URL: http://arxiv.org/abs/2207.09068v2
- Date: Wed, 20 Jul 2022 03:52:56 GMT
- Title: PiC: A Phrase-in-Context Dataset for Phrase Understanding and Semantic
Search
- Authors: Thang M. Pham, Seunghyun Yoon, Trung Bui, Anh Nguyen
- Abstract summary: We propose PiC - a dataset of 28K of noun phrases accompanied by their contextual Wikipedia pages.
We find that training on our dataset improves ranking models' accuracy and remarkably pushes Question Answering (QA) models to near-human accuracy.
- Score: 25.801066428860242
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Since BERT (Devlin et al., 2018), learning contextualized word embeddings has
been a de-facto standard in NLP. However, the progress of learning
contextualized phrase embeddings is hindered by the lack of a human-annotated,
phrase-in-context benchmark. To fill this gap, we propose PiC - a dataset of
~28K of noun phrases accompanied by their contextual Wikipedia pages and a
suite of three tasks of increasing difficulty for evaluating the quality of
phrase embeddings. We find that training on our dataset improves ranking
models' accuracy and remarkably pushes Question Answering (QA) models to
near-human accuracy which is 95% Exact Match (EM) on semantic search given a
query phrase and a passage. Interestingly, we find evidence that such
impressive performance is because the QA models learn to better capture the
common meaning of a phrase regardless of its actual context. That is, on our
Phrase Sense Disambiguation (PSD) task, SotA model accuracy drops substantially
(60% EM), failing to differentiate between two different senses of the same
phrase under two different contexts. Further results on our 3-task PiC
benchmark reveal that learning contextualized phrase embeddings remains an
interesting, open challenge.
Related papers
- Cross-lingual Contextualized Phrase Retrieval [63.80154430930898]
We propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval.
We train our Cross-lingual Contextualized Phrase Retriever (CCPR) using contrastive learning.
On the phrase retrieval task, CCPR surpasses baselines by a significant margin, achieving a top-1 accuracy that is at least 13 points higher.
arXiv Detail & Related papers (2024-03-25T14:46:51Z) - We're Afraid Language Models Aren't Modeling Ambiguity [136.8068419824318]
Managing ambiguity is a key part of human language understanding.
We characterize ambiguity in a sentence by its effect on entailment relations with another sentence.
We show that a multilabel NLI model can flag political claims in the wild that are misleading due to ambiguity.
arXiv Detail & Related papers (2023-04-27T17:57:58Z) - Approximate Nearest Neighbour Phrase Mining for Contextual Speech
Recognition [5.54562323810107]
We train end-to-end Context-Aware Transformer Transducer ( CATT) models by using a simple, yet efficient method of mining hard negative phrases from the latent space of the context encoder.
During training, given a reference query, we mine a number of similar phrases using approximate nearest neighbour search.
These sampled phrases are then used as negative examples in the context list alongside random and ground truth contextual information.
arXiv Detail & Related papers (2023-04-18T09:52:11Z) - Connect-the-Dots: Bridging Semantics between Words and Definitions via
Aligning Word Sense Inventories [47.03271152494389]
Word Sense Disambiguation aims to automatically identify the exact meaning of one word according to its context.
Existing supervised models struggle to make correct predictions on rare word senses due to limited training data.
We propose a gloss alignment algorithm that can align definition sentences with the same meaning from different sense inventories to collect rich lexical knowledge.
arXiv Detail & Related papers (2021-10-27T00:04:33Z) - UCPhrase: Unsupervised Context-aware Quality Phrase Tagging [63.86606855524567]
UCPhrase is a novel unsupervised context-aware quality phrase tagger.
We induce high-quality phrase spans as silver labels from consistently co-occurring word sequences.
We show that our design is superior to state-of-the-art pre-trained, unsupervised, and distantly supervised methods.
arXiv Detail & Related papers (2021-05-28T19:44:24Z) - Do Context-Aware Translation Models Pay the Right Attention? [61.25804242929533]
Context-aware machine translation models are designed to leverage contextual information, but often fail to do so.
In this paper, we ask several questions: What contexts do human translators use to resolve ambiguous words?
We introduce SCAT (Supporting Context for Ambiguous Translations), a new English-French dataset comprising supporting context words for 14K translations.
Using SCAT, we perform an in-depth analysis of the context used to disambiguate, examining positional and lexical characteristics of the supporting words.
arXiv Detail & Related papers (2021-05-14T17:32:24Z) - Non-Parametric Few-Shot Learning for Word Sense Disambiguation [11.175893018731712]
MetricWSD is a non-parametric few-shot learning approach to mitigate this data imbalance issue.
By learning to compute distances among the senses of a given word through episodic training, MetricWSD transfers knowledge from high-frequency words to infrequent ones.
arXiv Detail & Related papers (2021-04-26T16:08:46Z) - Pareto Probing: Trading Off Accuracy for Complexity [87.09294772742737]
We argue for a probe metric that reflects the fundamental trade-off between probe complexity and performance.
Our experiments with dependency parsing reveal a wide gap in syntactic knowledge between contextual and non-contextual representations.
arXiv Detail & Related papers (2020-10-05T17:27:31Z) - Fast and Robust Unsupervised Contextual Biasing for Speech Recognition [16.557586847398778]
We propose an alternative approach that does not entail explicit contextual language model.
We derive the bias score for every word in the system vocabulary from the training corpus.
We show significant improvement in recognition accuracy when the relevant context is available.
arXiv Detail & Related papers (2020-05-04T17:29:59Z) - Words aren't enough, their order matters: On the Robustness of Grounding
Visual Referring Expressions [87.33156149634392]
We critically examine RefCOg, a standard benchmark for visual referring expression recognition.
We show that 83.7% of test instances do not require reasoning on linguistic structure.
We propose two methods, one based on contrastive learning and the other based on multi-task learning, to increase the robustness of ViLBERT.
arXiv Detail & Related papers (2020-05-04T17:09:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.