Counterspeeches up my sleeve! Intent Distribution Learning and
Persistent Fusion for Intent-Conditioned Counterspeech Generation
- URL: http://arxiv.org/abs/2305.13776v1
- Date: Tue, 23 May 2023 07:45:17 GMT
- Title: Counterspeeches up my sleeve! Intent Distribution Learning and
Persistent Fusion for Intent-Conditioned Counterspeech Generation
- Authors: Rishabh Gupta, Shaily Desai, Manvi Goel, Anil Bandhakavi, Tanmoy
Chakraborty and Md. Shad Akhtar
- Abstract summary: In this paper, we explore intent-conditioned counterspeech generation.
We develop IntentCONAN, a diversified intent-specific counterspeech dataset with 6831 counterspeeches conditioned on five intents.
We propose QUARC, a two-stage framework for intent-conditioned counterspeech generation.
- Score: 26.4510688070963
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Counterspeech has been demonstrated to be an efficacious approach for
combating hate speech. While various conventional and controlled approaches
have been studied in recent years to generate counterspeech, a counterspeech
with a certain intent may not be sufficient in every scenario. Due to the
complex and multifaceted nature of hate speech, utilizing multiple forms of
counter-narratives with varying intents may be advantageous in different
circumstances. In this paper, we explore intent-conditioned counterspeech
generation. At first, we develop IntentCONAN, a diversified intent-specific
counterspeech dataset with 6831 counterspeeches conditioned on five intents,
i.e., informative, denouncing, question, positive, and humour. Subsequently, we
propose QUARC, a two-stage framework for intent-conditioned counterspeech
generation. QUARC leverages vector-quantized representations learned for each
intent category along with PerFuMe, a novel fusion module to incorporate
intent-specific information into the model. Our evaluation demonstrates that
QUARC outperforms several baselines by an average of 10% across evaluation
metrics. An extensive human evaluation supplements our hypothesis of better and
more appropriate responses than comparative systems.
Related papers
- CrowdCounter: A benchmark type-specific multi-target counterspeech dataset [10.133642589954192]
We introduce a new dataset - CrowdCounter containing 3,425 hate speech-counterspeech pairs.
The design of our annotation platform itself encourages annotators to write type-specific, non-redundant and high-quality counterspeech.
We evaluate two frameworks for generating counterspeech responses - vanilla and type-controlled prompts.
arXiv Detail & Related papers (2024-10-02T10:24:51Z) - SpeechAlign: Aligning Speech Generation to Human Preferences [51.684183257809075]
We introduce SpeechAlign, an iterative self-improvement strategy that aligns speech language models to human preferences.
We show that SpeechAlign can bridge the distribution gap and facilitate continuous self-improvement of the speech language model.
arXiv Detail & Related papers (2024-04-08T15:21:17Z) - Outcome-Constrained Large Language Models for Countering Hate Speech [10.434435022492723]
This study aims to develop methods for generating counterspeech constrained by conversation outcomes.
We experiment with large language models (LLMs) to incorporate into the text generation process two desired conversation outcomes.
Evaluation results show that our methods effectively steer the generation of counterspeech toward the desired outcomes.
arXiv Detail & Related papers (2024-03-25T19:44:06Z) - Intent-conditioned and Non-toxic Counterspeech Generation using Multi-Task Instruction Tuning with RLAIF [14.2594830589926]
Counterspeech, defined as a response to online hate speech, is increasingly used as a non-censorial solution.
Our study introduces CoARL, a novel framework enhancing counterspeech generation by modeling the pragmatic implications underlying social biases in hateful statements.
CoARL's first two phases involve sequential multi-instruction tuning, teaching the model to understand intents, reactions, and harms of offensive statements, and then learning task-specific low-rank adapter weights for generating intent-conditioned counterspeech.
arXiv Detail & Related papers (2024-03-15T08:03:49Z) - DisCGen: A Framework for Discourse-Informed Counterspeech Generation [34.75404551612012]
We propose a framework based on theories of discourse to study the inferential links that connect counter speeches to hateful comments.
We present a process for collecting an in-the-wild dataset of counterspeech from Reddit.
We show that by using our dataset and framework, large language models can generate contextually-grounded counterspeech informed by theories of discourse.
arXiv Detail & Related papers (2023-11-29T23:20:17Z) - Exploring Speech Recognition, Translation, and Understanding with
Discrete Speech Units: A Comparative Study [68.88536866933038]
Speech signals, typically sampled at rates in the tens of thousands per second, contain redundancies.
Recent investigations proposed the use of discrete speech units derived from self-supervised learning representations.
Applying various methods, such as de-duplication and subword modeling, can further compress the speech sequence length.
arXiv Detail & Related papers (2023-09-27T17:21:13Z) - Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of
Language Models [86.02610674750345]
Adversarial GLUE (AdvGLUE) is a new multi-task benchmark to explore and evaluate the vulnerabilities of modern large-scale language models under various types of adversarial attacks.
We apply 14 adversarial attack methods to GLUE tasks to construct AdvGLUE, which is further validated by humans for reliable annotations.
All the language models and robust training methods we tested perform poorly on AdvGLUE, with scores lagging far behind the benign accuracy.
arXiv Detail & Related papers (2021-11-04T12:59:55Z) - Fuzzy Classification of Multi-intent Utterances [0.0]
Current intent classification approaches assign binary intent class memberships to natural language utterances.
We propose a scheme to address the ambiguity in single-intent as well as multi-intent natural language utterances.
arXiv Detail & Related papers (2021-04-22T02:15:56Z) - General-Purpose Speech Representation Learning through a Self-Supervised
Multi-Granularity Framework [114.63823178097402]
This paper presents a self-supervised learning framework, named MGF, for general-purpose speech representation learning.
Specifically, we propose to use generative learning approaches to capture fine-grained information at small time scales and use discriminative learning approaches to distill coarse-grained or semantic information at large time scales.
arXiv Detail & Related papers (2021-02-03T08:13:21Z) - Discriminative Nearest Neighbor Few-Shot Intent Detection by
Transferring Natural Language Inference [150.07326223077405]
Few-shot learning is attracting much attention to mitigate data scarcity.
We present a discriminative nearest neighbor classification with deep self-attention.
We propose to boost the discriminative ability by transferring a natural language inference (NLI) model.
arXiv Detail & Related papers (2020-10-25T00:39:32Z) - Unsupervised Cross-lingual Representation Learning for Speech
Recognition [63.85924123692923]
XLSR learns cross-lingual speech representations by pretraining a single model from the raw waveform of speech in multiple languages.
We build on wav2vec 2.0 which is trained by solving a contrastive task over masked latent speech representations.
Experiments show that cross-lingual pretraining significantly outperforms monolingual pretraining.
arXiv Detail & Related papers (2020-06-24T18:25:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.