NeuroCounterfactuals: Beyond Minimal-Edit Counterfactuals for Richer
Data Augmentation
- URL: http://arxiv.org/abs/2210.12365v1
- Date: Sat, 22 Oct 2022 06:29:21 GMT
- Title: NeuroCounterfactuals: Beyond Minimal-Edit Counterfactuals for Richer
Data Augmentation
- Authors: Phillip Howard, Gadi Singer, Vasudev Lal, Yejin Choi, Swabha
Swayamdipta
- Abstract summary: We introduce NeuroCounterfactuals, designed as loose counterfactuals, allowing for larger edits which result in naturalistic generations containing linguistic diversity.
Our novel generative approach bridges the benefits of constrained decoding, with those of language model adaptation for sentiment steering.
- Score: 55.17069935305069
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While counterfactual data augmentation offers a promising step towards robust
generalization in natural language processing, producing a set of
counterfactuals that offer valuable inductive bias for models remains a
challenge. Most existing approaches for producing counterfactuals, manual or
automated, rely on small perturbations via minimal edits, resulting in
simplistic changes. We introduce NeuroCounterfactuals, designed as loose
counterfactuals, allowing for larger edits which result in naturalistic
generations containing linguistic diversity, while still bearing similarity to
the original document. Our novel generative approach bridges the benefits of
constrained decoding, with those of language model adaptation for sentiment
steering. Training data augmentation with our generations results in both
in-domain and out-of-domain improvements for sentiment classification,
outperforming even manually curated counterfactuals, under select settings. We
further present detailed analyses to show the advantages of
NeuroCounterfactuals over approaches involving simple, minimal edits.
Related papers
- Contextual Biasing to Improve Domain-specific Custom Vocabulary Audio Transcription without Explicit Fine-Tuning of Whisper Model [0.0]
OpenAI's Whisper Automated Speech Recognition model excels in generalizing across diverse datasets and domains.
We propose a method to enhance transcription accuracy without explicit fine-tuning or altering model parameters.
arXiv Detail & Related papers (2024-10-24T01:58:11Z) - HyPoradise: An Open Baseline for Generative Speech Recognition with
Large Language Models [81.56455625624041]
We introduce the first open-source benchmark to utilize external large language models (LLMs) for ASR error correction.
The proposed benchmark contains a novel dataset, HyPoradise (HP), encompassing more than 334,000 pairs of N-best hypotheses.
LLMs with reasonable prompt and its generative capability can even correct those tokens that are missing in N-best list.
arXiv Detail & Related papers (2023-09-27T14:44:10Z) - DiffusER: Discrete Diffusion via Edit-based Reconstruction [88.62707047517914]
DiffusER is an edit-based generative model for text based on denoising diffusion models.
It can rival autoregressive models on several tasks spanning machine translation, summarization, and style transfer.
It can also perform other varieties of generation that standard autoregressive models are not well-suited for.
arXiv Detail & Related papers (2022-10-30T16:55:23Z) - CORE: A Retrieve-then-Edit Framework for Counterfactual Data Generation [91.16551253297588]
COunterfactual Generation via Retrieval and Editing (CORE) is a retrieval-augmented generation framework for creating diverse counterfactual perturbations for training.
CORE first performs a dense retrieval over a task-related unlabeled text corpus using a learned bi-encoder.
CORE then incorporates these into prompts to a large language model with few-shot learning capabilities, for counterfactual editing.
arXiv Detail & Related papers (2022-10-10T17:45:38Z) - Factorized Neural Transducer for Efficient Language Model Adaptation [51.81097243306204]
We propose a novel model, factorized neural Transducer, by factorizing the blank and vocabulary prediction.
It is expected that this factorization can transfer the improvement of the standalone language model to the Transducer for speech recognition.
We demonstrate that the proposed factorized neural Transducer yields 15% to 20% WER improvements when out-of-domain text data is used for language model adaptation.
arXiv Detail & Related papers (2021-09-27T15:04:00Z) - Learning Neural Models for Natural Language Processing in the Face of
Distributional Shift [10.990447273771592]
The dominating NLP paradigm of training a strong neural predictor to perform one task on a specific dataset has led to state-of-the-art performance in a variety of applications.
It builds upon the assumption that the data distribution is stationary, ie. that the data is sampled from a fixed distribution both at training and test time.
This way of training is inconsistent with how we as humans are able to learn from and operate within a constantly changing stream of information.
It is ill-adapted to real-world use cases where the data distribution is expected to shift over the course of a model's lifetime
arXiv Detail & Related papers (2021-09-03T14:29:20Z) - End-to-end Neural Coreference Resolution Revisited: A Simple yet
Effective Baseline [20.431647446999996]
We propose a simple yet effective baseline for coreference resolution.
Our model is a simplified version of the original neural coreference resolution model.
Our work provides evidence for the necessity of carefully justifying the complexity of existing or newly proposed models.
arXiv Detail & Related papers (2021-07-04T18:12:24Z) - SDA: Improving Text Generation with Self Data Augmentation [88.24594090105899]
We propose to improve the standard maximum likelihood estimation (MLE) paradigm by incorporating a self-imitation-learning phase for automatic data augmentation.
Unlike most existing sentence-level augmentation strategies, our method is more general and could be easily adapted to any MLE-based training procedure.
arXiv Detail & Related papers (2021-01-02T01:15:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.