APPDIA: A Discourse-aware Transformer-based Style Transfer Model for
Offensive Social Media Conversations
- URL: http://arxiv.org/abs/2209.08207v1
- Date: Sat, 17 Sep 2022 00:50:24 GMT
- Title: APPDIA: A Discourse-aware Transformer-based Style Transfer Model for
Offensive Social Media Conversations
- Authors: Katherine Atwell, Sabit Hassan, Malihe Alikhani
- Abstract summary: We release the first publicly-available, parallel corpus of offensive Reddit comments and their style-transferred counterparts annotated by sociolinguists.
We introduce the first discourse-aware style-transfer models that can effectively reduce offensiveness in Reddit text while preserving the meaning of the original text.
- Score: 11.011242089340437
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Using style-transfer models to reduce offensiveness of social media comments
can help foster a more inclusive environment. However, there are no sizable
datasets that contain offensive texts and their inoffensive counterparts, and
fine-tuning pretrained models with limited labeled data can lead to the loss of
original meaning in the style-transferred text. To address this issue, we
provide two major contributions. First, we release the first
publicly-available, parallel corpus of offensive Reddit comments and their
style-transferred counterparts annotated by expert sociolinguists. Then, we
introduce the first discourse-aware style-transfer models that can effectively
reduce offensiveness in Reddit text while preserving the meaning of the
original text. These models are the first to examine inferential links between
the comment and the text it is replying to when transferring the style of
offensive Reddit text. We propose two different methods of integrating
discourse relations with pretrained transformer models and evaluate them on our
dataset of offensive comments from Reddit and their inoffensive counterparts.
Improvements over the baseline with respect to both automatic metrics and human
evaluation indicate that our discourse-aware models are better at preserving
meaning in style-transferred text when compared to the state-of-the-art
discourse-agnostic models.
Related papers
- Prefix-Tuning Based Unsupervised Text Style Transfer [29.86587278794342]
Unsupervised text style transfer aims at training a generative model that can alter the style of the input sentence while preserving its content.
In this paper, we employ powerful pre-trained large language models and present a new prefix-tuning-based method for unsupervised text style transfer.
arXiv Detail & Related papers (2023-10-23T06:13:08Z) - Understanding writing style in social media with a supervised
contrastively pre-trained transformer [57.48690310135374]
Online Social Networks serve as fertile ground for harmful behavior, ranging from hate speech to the dissemination of disinformation.
We introduce the Style Transformer for Authorship Representations (STAR), trained on a large corpus derived from public sources of 4.5 x 106 authored texts.
Using a support base of 8 documents of 512 tokens, we can discern authors from sets of up to 1616 authors with at least 80% accuracy.
arXiv Detail & Related papers (2023-10-17T09:01:17Z) - Specializing Small Language Models towards Complex Style Transfer via
Latent Attribute Pre-Training [29.143887057933327]
We introduce the concept of complex text style transfer tasks, and constructed complex text datasets based on two widely applicable scenarios.
Our dataset is the first large-scale data set of its kind, with 700 rephrased sentences and 1,000 sentences from the game Genshin Impact.
arXiv Detail & Related papers (2023-09-19T21:01:40Z) - ParaGuide: Guided Diffusion Paraphrasers for Plug-and-Play Textual Style
Transfer [57.6482608202409]
Textual style transfer is the task of transforming stylistic properties of text while preserving meaning.
We introduce a novel diffusion-based framework for general-purpose style transfer that can be flexibly adapted to arbitrary target styles.
We validate the method on the Enron Email Corpus, with both human and automatic evaluations, and find that it outperforms strong baselines on formality, sentiment, and even authorship style transfer.
arXiv Detail & Related papers (2023-08-29T17:36:02Z) - Text Detoxification using Large Pre-trained Neural Models [57.72086777177844]
We present two novel unsupervised methods for eliminating toxicity in text.
First method combines guidance of the generation process with small style-conditional language models.
Second method uses BERT to replace toxic words with their non-offensive synonyms.
arXiv Detail & Related papers (2021-09-18T11:55:32Z) - GTAE: Graph-Transformer based Auto-Encoders for Linguistic-Constrained
Text Style Transfer [119.70961704127157]
Non-parallel text style transfer has attracted increasing research interests in recent years.
Current approaches still lack the ability to preserve the content and even logic of original sentences.
We propose a method called Graph Transformer based Auto-GTAE, which models a sentence as a linguistic graph and performs feature extraction and style transfer at the graph level.
arXiv Detail & Related papers (2021-02-01T11:08:45Z) - Towards A Friendly Online Community: An Unsupervised Style Transfer
Framework for Profanity Redaction [17.380204095038795]
Offensive and abusive language is a pressing problem on social media platforms.
We propose a method for transforming offensive comments, statements containing profanity or offensive language, into non-offensive ones.
arXiv Detail & Related papers (2020-11-01T02:10:25Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - Contextual Text Style Transfer [73.66285813595616]
Contextual Text Style Transfer aims to translate a sentence into a desired style with its surrounding context taken into account.
We propose a Context-Aware Style Transfer (CAST) model, which uses two separate encoders for each input sentence and its surrounding context.
Two new benchmarks, Enron-Context and Reddit-Context, are introduced for formality and offensiveness style transfer.
arXiv Detail & Related papers (2020-04-30T23:01:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.