Less Likely Brainstorming: Using Language Models to Generate Alternative
Hypotheses
- URL: http://arxiv.org/abs/2305.19339v1
- Date: Tue, 30 May 2023 18:05:34 GMT
- Title: Less Likely Brainstorming: Using Language Models to Generate Alternative
Hypotheses
- Authors: Liyan Tang, Yifan Peng, Yanshan Wang, Ying Ding, Greg Durrett, Justin
F. Rousseau
- Abstract summary: We introduce a new task, "less likely brainstorming," that asks a model to generate outputs that humans think are relevant but less likely to happen.
We find that a baseline approach of training with less likely hypotheses as targets generates outputs that humans evaluate as either likely or irrelevant nearly half of the time.
We propose a controlled text generation method that uses a novel contrastive learning strategy to encourage models to differentiate between generating likely and less likely outputs according to humans.
- Score: 45.720065723998225
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A human decision-maker benefits the most from an AI assistant that corrects
for their biases. For problems such as generating interpretation of a radiology
report given findings, a system predicting only highly likely outcomes may be
less useful, where such outcomes are already obvious to the user. To alleviate
biases in human decision-making, it is worth considering a broad differential
diagnosis, going beyond the most likely options. We introduce a new task, "less
likely brainstorming," that asks a model to generate outputs that humans think
are relevant but less likely to happen. We explore the task in two settings: a
brain MRI interpretation generation setting and an everyday commonsense
reasoning setting. We found that a baseline approach of training with less
likely hypotheses as targets generates outputs that humans evaluate as either
likely or irrelevant nearly half of the time; standard MLE training is not
effective. To tackle this problem, we propose a controlled text generation
method that uses a novel contrastive learning strategy to encourage models to
differentiate between generating likely and less likely outputs according to
humans. We compare our method with several state-of-the-art controlled text
generation models via automatic and human evaluations and show that our models'
capability of generating less likely outputs is improved.
Related papers
- How Aligned are Generative Models to Humans in High-Stakes Decision-Making? [10.225573060836478]
Large generative models (LMs) are increasingly being considered for high-stakes decision-making.
This work considers how such models compare to humans and predictive AI models on a specific case of recidivism prediction.
arXiv Detail & Related papers (2024-10-20T19:00:59Z) - Self-Debiasing Large Language Models: Zero-Shot Recognition and
Reduction of Stereotypes [73.12947922129261]
We leverage the zero-shot capabilities of large language models to reduce stereotyping.
We show that self-debiasing can significantly reduce the degree of stereotyping across nine different social groups.
We hope this work opens inquiry into other zero-shot techniques for bias mitigation.
arXiv Detail & Related papers (2024-02-03T01:40:11Z) - Mitigating Biases for Instruction-following Language Models via Bias Neurons Elimination [54.865941973768905]
We propose a novel and practical bias mitigation method, CRISPR, to eliminate bias neurons of language models in instruction-following settings.
CRISPR automatically determines biased outputs and categorizes neurons that affect the biased outputs as bias neurons using an explainability method.
Experimental results demonstrate the effectiveness of our method in mitigating biases under zero-shot instruction-following settings without losing the model's task performance and existing knowledge.
arXiv Detail & Related papers (2023-11-16T07:16:55Z) - Fine-tuning Language Models for Factuality [96.5203774943198]
Large pre-trained language models (LLMs) have led to their widespread use, sometimes even as a replacement for traditional search engines.
Yet language models are prone to making convincing but factually inaccurate claims, often referred to as 'hallucinations'
In this work, we fine-tune language models to be more factual, without human labeling.
arXiv Detail & Related papers (2023-11-14T18:59:15Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - The Boltzmann Policy Distribution: Accounting for Systematic
Suboptimality in Human Models [5.736353542430439]
We introduce the Boltzmann policy distribution (BPD), which serves as a prior over human policies.
BPD adapts via Bayesian inference to capture systematic deviations by observing human actions during a single episode.
We show that the BPD enables prediction of human behavior and human-AI collaboration equally as well as imitation learning-based human models.
arXiv Detail & Related papers (2022-04-22T15:26:25Z) - Typical Decoding for Natural Language Generation [76.69397802617064]
We study why high-probability texts can be dull or repetitive.
We show that typical sampling offers competitive performance in terms of quality.
arXiv Detail & Related papers (2022-02-01T18:58:45Z) - Humans learn too: Better Human-AI Interaction using Optimized Human
Inputs [2.5991265608180396]
Humans rely more and more on systems with AI components.
The AI community typically treats human inputs as a given and optimize AI models only.
In this work, human inputs are optimized for better interaction with an AI model while keeping the model fixed.
arXiv Detail & Related papers (2020-09-19T16:30:37Z) - Classification Under Human Assistance [29.220005688025378]
We show that supervised learning models trained to operate under different automation levels can outperform those trained for full automation as well as humans operating alone.
Experiments on synthetic and real-world data from several applications in medical diagnosis illustrate our theoretical findings.
arXiv Detail & Related papers (2020-06-21T16:52:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.