When Hearst Is not Enough: Improving Hypernymy Detection from Corpus
with Distributional Models
- URL: http://arxiv.org/abs/2010.04941v1
- Date: Sat, 10 Oct 2020 08:34:19 GMT
- Title: When Hearst Is not Enough: Improving Hypernymy Detection from Corpus
with Distributional Models
- Authors: Changlong Yu, Jialong Han, Peifeng Wang, Yangqiu Song, Hongming Zhang,
Wilfred Ng, Shuming Shi
- Abstract summary: This paper addresses whether an is-a relationship exists between words (x, y) with the help of large textual corpora.
Recent studies suggest that pattern-based ones are superior, if large-scale Hearst pairs are extracted and fed, with the sparsity of unseen (x, y) pairs relieved.
For the first time, this paper quantifies the non-negligible existence of those specific cases. We also demonstrate that distributional methods are ideal to make up for pattern-based ones in such cases.
- Score: 59.46552488974247
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We address hypernymy detection, i.e., whether an is-a relationship exists
between words (x, y), with the help of large textual corpora. Most conventional
approaches to this task have been categorized to be either pattern-based or
distributional. Recent studies suggest that pattern-based ones are superior, if
large-scale Hearst pairs are extracted and fed, with the sparsity of unseen (x,
y) pairs relieved. However, they become invalid in some specific sparsity
cases, where x or y is not involved in any pattern. For the first time, this
paper quantifies the non-negligible existence of those specific cases. We also
demonstrate that distributional methods are ideal to make up for pattern-based
ones in such cases. We devise a complementary framework, under which a
pattern-based and a distributional model collaborate seamlessly in cases which
they each prefer. On several benchmark datasets, our framework achieves
competitive improvements and the case study shows its better interpretability.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Prototype-based Aleatoric Uncertainty Quantification for Cross-modal
Retrieval [139.21955930418815]
Cross-modal Retrieval methods build similarity relations between vision and language modalities by jointly learning a common representation space.
However, the predictions are often unreliable due to the Aleatoric uncertainty, which is induced by low-quality data, e.g., corrupt images, fast-paced videos, and non-detailed texts.
We propose a novel Prototype-based Aleatoric Uncertainty Quantification (PAU) framework to provide trustworthy predictions by quantifying the uncertainty arisen from the inherent data ambiguity.
arXiv Detail & Related papers (2023-09-29T09:41:19Z) - Investigating Ensemble Methods for Model Robustness Improvement of Text
Classifiers [66.36045164286854]
We analyze a set of existing bias features and demonstrate there is no single model that works best for all the cases.
By choosing an appropriate bias model, we can obtain a better robustness result than baselines with a more sophisticated model design.
arXiv Detail & Related papers (2022-10-28T17:52:10Z) - GAPX: Generalized Autoregressive Paraphrase-Identification X [24.331570697458954]
A major source of this performance drop comes from biases introduced by negative examples.
We introduce a perplexity based out-of-distribution metric that we show can effectively and automatically determine how much weight it should be given during inference.
arXiv Detail & Related papers (2022-10-05T01:23:52Z) - Uncertainty Estimation for Language Reward Models [5.33024001730262]
Language models can learn a range of capabilities from unsupervised training on text corpora.
It is often easier for humans to choose between options than to provide labeled data, and prior work has achieved state-of-the-art performance by training a reward model from such preference comparisons.
We seek to address these problems via uncertainty estimation, which can improve sample efficiency and robustness using active learning and risk-averse reinforcement learning.
arXiv Detail & Related papers (2022-03-14T20:13:21Z) - Finding and Fixing Spurious Patterns with Explanations [14.591545536354621]
We present an end-to-end pipeline for identifying and mitigating spurious patterns for image classifiers.
We find patterns such as "the model's prediction for tennis racket changes 63% of the time if we hide the people"
Then, if a pattern is spurious, we mitigate it via a novel form of data augmentation.
arXiv Detail & Related papers (2021-06-03T20:07:46Z) - An Effective Baseline for Robustness to Distributional Shift [5.627346969563955]
Refraining from confidently predicting when faced with categories of inputs different from those seen during training is an important requirement for the safe deployment of deep learning systems.
We present a simple, but highly effective approach to deal with out-of-distribution detection that uses the principle of abstention.
arXiv Detail & Related papers (2021-05-15T00:46:11Z) - Unsupervised Anomaly Detection From Semantic Similarity Scores [0.0]
We present a simple and generic framework, it SemSAD, that makes use of a semantic similarity score to carry out anomaly detection.
We are able to outperform previous approaches for anomaly, novelty, or out-of-distribution detection in the visual domain by a large margin.
arXiv Detail & Related papers (2020-12-01T13:12:31Z) - Squared $\ell_2$ Norm as Consistency Loss for Leveraging Augmented Data
to Learn Robust and Invariant Representations [76.85274970052762]
Regularizing distance between embeddings/representations of original samples and augmented counterparts is a popular technique for improving robustness of neural networks.
In this paper, we explore these various regularization choices, seeking to provide a general understanding of how we should regularize the embeddings.
We show that the generic approach we identified (squared $ell$ regularized augmentation) outperforms several recent methods, which are each specially designed for one task.
arXiv Detail & Related papers (2020-11-25T22:40:09Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.