Distributed NLI: Learning to Predict Human Opinion Distributions for
Language Reasoning
- URL: http://arxiv.org/abs/2104.08676v1
- Date: Sun, 18 Apr 2021 01:25:19 GMT
- Title: Distributed NLI: Learning to Predict Human Opinion Distributions for
Language Reasoning
- Authors: Xiang Zhou, Yixin Nie, Mohit Bansal
- Abstract summary: We introduce distributed NLI, a new NLU task with a goal to predict the distribution of human judgements for natural language inference.
We show that models can capture human judgement distribution by applying additional distribution estimation methods, namely, Monte Carlo (MC) Dropout, Deep Ensemble, Re-Calibration, and Distribution Distillation.
- Score: 76.17436599516074
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce distributed NLI, a new NLU task with a goal to predict the
distribution of human judgements for natural language inference. We show that
models can capture human judgement distribution by applying additional
distribution estimation methods, namely, Monte Carlo (MC) Dropout, Deep
Ensemble, Re-Calibration, and Distribution Distillation. All four of these
methods substantially outperform the softmax baseline. We show that MC Dropout
is able to achieve decent performance without any distribution annotations
while Re-Calibration can further give substantial improvements when extra
distribution annotations are provided, suggesting the value of multiple
annotations for the example in modeling the distribution of human judgements.
Moreover, MC Dropout and Re-Calibration can achieve decent transfer performance
on out-of-domain data. Despite these improvements, the best results are still
far below estimated human upper-bound, indicating that the task of predicting
the distribution of human judgements is still an open, challenging problem with
large room for future improvements. We showcase the common errors for MC
Dropout and Re-Calibration. Finally, we give guidelines on the usage of these
methods with different levels of data availability and encourage future work on
modeling the human opinion distribution for language reasoning.
Related papers
- Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Personalized Trajectory Prediction via Distribution Discrimination [78.69458579657189]
Trarimiy prediction is confronted with the dilemma to capture the multi-modal nature of future dynamics.
We present a distribution discrimination (DisDis) method to predict personalized motion patterns.
Our method can be integrated with existing multi-modal predictive models as a plug-and-play module.
arXiv Detail & Related papers (2021-07-29T17:42:12Z) - Predicting with Confidence on Unseen Distributions [90.68414180153897]
We connect domain adaptation and predictive uncertainty literature to predict model accuracy on challenging unseen distributions.
We find that the difference of confidences (DoC) of a classifier's predictions successfully estimates the classifier's performance change over a variety of shifts.
We specifically investigate the distinction between synthetic and natural distribution shifts and observe that despite its simplicity DoC consistently outperforms other quantifications of distributional difference.
arXiv Detail & Related papers (2021-07-07T15:50:18Z) - Mandoline: Model Evaluation under Distribution Shift [8.007644303175395]
Machine learning models are often deployed in different settings than they were trained and validated on.
We develop Mandoline, a new evaluation framework that mitigates these issues.
Users write simple "slicing functions" - noisy, potentially correlated binary functions intended to capture possible axes of distribution shift.
arXiv Detail & Related papers (2021-07-01T17:57:57Z) - Distributional Reinforcement Learning via Moment Matching [54.16108052278444]
We formulate a method that learns a finite set of statistics from each return distribution via neural networks.
Our method can be interpreted as implicitly matching all orders of moments between a return distribution and its Bellman target.
Experiments on the suite of Atari games show that our method outperforms the standard distributional RL baselines.
arXiv Detail & Related papers (2020-07-24T05:18:17Z) - Mind the Trade-off: Debiasing NLU Models without Degrading the
In-distribution Performance [70.31427277842239]
We introduce a novel debiasing method called confidence regularization.
It discourages models from exploiting biases while enabling them to receive enough incentive to learn from all the training examples.
We evaluate our method on three NLU tasks and show that, in contrast to its predecessors, it improves the performance on out-of-distribution datasets.
arXiv Detail & Related papers (2020-05-01T11:22:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.