Towards Robust NLG Bias Evaluation with Syntactically-diverse Prompts
- URL: http://arxiv.org/abs/2212.01700v1
- Date: Sat, 3 Dec 2022 22:11:17 GMT
- Title: Towards Robust NLG Bias Evaluation with Syntactically-diverse Prompts
- Authors: Arshiya Aggarwal, Jiao Sun, Nanyun Peng
- Abstract summary: We present a robust methodology for evaluating biases in natural language generation (NLG) systems.
Previous works use fixed hand-crafted prefix templates with mentions of various demographic groups to prompt models to generate continuations for bias analysis.
To study this problem, we paraphrase the prompts with different syntactic structures and use these to evaluate demographic bias in NLG systems.
- Score: 38.69716232707304
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We present a robust methodology for evaluating biases in natural language
generation(NLG) systems. Previous works use fixed hand-crafted prefix templates
with mentions of various demographic groups to prompt models to generate
continuations for bias analysis. These fixed prefix templates could themselves
be specific in terms of styles or linguistic structures, which may lead to
unreliable fairness conclusions that are not representative of the general
trends from tone varying prompts. To study this problem, we paraphrase the
prompts with different syntactic structures and use these to evaluate
demographic bias in NLG systems. Our results suggest similar overall bias
trends but some syntactic structures lead to contradictory conclusions compared
to past works. We show that our methodology is more robust and that some
syntactic structures prompt more toxic content while others could prompt less
biased generation. This suggests the importance of not relying on a fixed
syntactic structure and using tone-invariant prompts. Introducing
syntactically-diverse prompts can achieve more robust NLG (bias) evaluation.
Related papers
- A Closer Look at Bias and Chain-of-Thought Faithfulness of Large (Vision) Language Models [53.18562650350898]
Chain-of-thought (CoT) reasoning enhances performance of large language models.<n>We present the first comprehensive study of CoT faithfulness in large vision-language models.
arXiv Detail & Related papers (2025-05-29T18:55:05Z) - SHINE: Saliency-aware HIerarchical NEgative Ranking for Compositional Temporal Grounding [52.98133831401225]
Temporal grounding, also known as video moment retrieval, aims at locating video segments corresponding to a given query sentence.
We propose a large language model-driven method for negative query construction, utilizing GPT-3.5-Turbo.
We introduce a coarse-to-fine saliency ranking strategy, which encourages the model to learn the multi-granularity semantic relationships between videos and hierarchical negative queries.
arXiv Detail & Related papers (2024-07-06T16:08:17Z) - Fairness-guided Few-shot Prompting for Large Language Models [93.05624064699965]
In-context learning can suffer from high instability due to variations in training examples, example order, and prompt formats.
We introduce a metric to evaluate the predictive bias of a fixed prompt against labels or a given attributes.
We propose a novel search strategy based on the greedy search to identify the near-optimal prompt for improving the performance of in-context learning.
arXiv Detail & Related papers (2023-03-23T12:28:25Z) - Language model acceptability judgements are not always robust to context [30.868765627701457]
We investigate the stability of language models' performance on targeted syntactic evaluations.
We find that model judgements are generally robust when placed in randomly sampled linguistic contexts.
We show that these changes in model performance are not explainable by simple features matching the context and the test inputs.
arXiv Detail & Related papers (2022-12-18T00:11:06Z) - Using Natural Sentences for Understanding Biases in Language Models [10.604991889372092]
We create a prompt dataset with respect to occupations collected from real-world natural sentences in Wikipedia.
We find bias evaluations are very sensitive to the design choices of template prompts.
We propose using natural sentence prompts for systematic evaluations to step away from design choices that could introduce bias in the observations.
arXiv Detail & Related papers (2022-05-12T18:36:33Z) - Does BERT really agree ? Fine-grained Analysis of Lexical Dependence on
a Syntactic Task [70.29624135819884]
We study the extent to which BERT is able to perform lexically-independent subject-verb number agreement (NA) on targeted syntactic templates.
Our results on nonce sentences suggest that the model generalizes well for simple templates, but fails to perform lexically-independent syntactic generalization when as little as one attractor is present.
arXiv Detail & Related papers (2022-04-14T11:33:15Z) - The Authors Matter: Understanding and Mitigating Implicit Bias in Deep
Text Classification [36.361778457307636]
Deep text classification models can produce biased outcomes for texts written by authors of certain demographic groups.
In this paper, we first demonstrate that implicit bias exists in different text classification tasks for different demographic groups.
We then build a learning-based interpretation method to deepen our knowledge of implicit bias.
arXiv Detail & Related papers (2021-05-06T16:17:38Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Improving Robustness by Augmenting Training Sentences with
Predicate-Argument Structures [62.562760228942054]
Existing approaches to improve robustness against dataset biases mostly focus on changing the training objective.
We propose to augment the input sentences in the training data with their corresponding predicate-argument structures.
We show that without targeting a specific bias, our sentence augmentation improves the robustness of transformer models against multiple biases.
arXiv Detail & Related papers (2020-10-23T16:22:05Z) - Recurrent Neural Network Language Models Always Learn English-Like
Relative Clause Attachment [17.995905582226463]
We compare model performance in English and Spanish to show that non-linguistic biases in RNN LMs advantageously overlap with syntactic structure in English but not Spanish.
English models may appear to acquire human-like syntactic preferences, while models trained on Spanish fail to acquire comparable human-like preferences.
arXiv Detail & Related papers (2020-05-01T01:21:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.