Aggressive Language Detection with Joint Text Normalization via
Adversarial Multi-task Learning
- URL: http://arxiv.org/abs/2009.09174v1
- Date: Sat, 19 Sep 2020 06:26:07 GMT
- Title: Aggressive Language Detection with Joint Text Normalization via
Adversarial Multi-task Learning
- Authors: Shengqiong Wu and Hao Fei and Donghong Ji
- Abstract summary: Aggressive language detection (ALD) is one of the crucial applications in NLP community.
In this work, we target improving the ALD by jointly performing text normalization (TN), via an adversarial multi-task learning framework.
- Score: 31.02484600391725
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Aggressive language detection (ALD), detecting the abusive and offensive
language in texts, is one of the crucial applications in NLP community. Most
existing works treat ALD as regular classification with neural models, while
ignoring the inherent conflicts of social media text that they are quite
unnormalized and irregular. In this work, we target improving the ALD by
jointly performing text normalization (TN), via an adversarial multi-task
learning framework. The private encoders for ALD and TN focus on the
task-specific features retrieving, respectively, and the shared encoder learns
the underlying common features over two tasks. During adversarial training, a
task discriminator distinguishes the separate learning of ALD or TN.
Experimental results on four ALD datasets show that our model outperforms all
baselines under differing settings by large margins, demonstrating the
necessity of joint learning the TN with ALD. Further analysis is conducted for
a better understanding of our method.
Related papers
- A Unified Multi-Task Learning Architecture for Hate Detection Leveraging User-Based Information [23.017068553977982]
Hate speech, offensive language, aggression, racism, sexism, and other abusive language are common phenomena in social media.
There is a need for Artificial Intelligence(AI)based intervention which can filter hate content at scale.
This paper introduces a unique model that improves hate speech identification for the English language by utilising intra-user and inter-user-based information.
arXiv Detail & Related papers (2024-11-11T10:37:11Z) - Aligning and Prompting Everything All at Once for Universal Visual
Perception [79.96124061108728]
APE is a universal visual perception model for aligning and prompting everything all at once in an image to perform diverse tasks.
APE advances the convergence of detection and grounding by reformulating language-guided grounding as open-vocabulary detection.
Experiments on over 160 datasets demonstrate that APE outperforms state-of-the-art models.
arXiv Detail & Related papers (2023-12-04T18:59:50Z) - Automatic Textual Normalization for Hate Speech Detection [0.8990550886501417]
Social media data contains a wide range of non-standard words (NSW)
Current state-of-the-art methods for the Vietnamese language address this issue as a problem of lexical normalization.
Our approach is straightforward, employing solely a sequence-to-sequence (Seq2Seq) model.
arXiv Detail & Related papers (2023-11-12T14:01:38Z) - Pre-training Multi-task Contrastive Learning Models for Scientific
Literature Understanding [52.723297744257536]
Pre-trained language models (LMs) have shown effectiveness in scientific literature understanding tasks.
We propose a multi-task contrastive learning framework, SciMult, to facilitate common knowledge sharing across different literature understanding tasks.
arXiv Detail & Related papers (2023-05-23T16:47:22Z) - Self-Supervised Graph Neural Network for Multi-Source Domain Adaptation [51.21190751266442]
Domain adaptation (DA) tries to tackle the scenarios when the test data does not fully follow the same distribution of the training data.
By learning from large-scale unlabeled samples, self-supervised learning has now become a new trend in deep learning.
We propose a novel textbfSelf-textbfSupervised textbfGraph Neural Network (SSG) to enable more effective inter-task information exchange and knowledge sharing.
arXiv Detail & Related papers (2022-04-08T03:37:56Z) - Improving Multi-task Generalization Ability for Neural Text Matching via
Prompt Learning [54.66399120084227]
Recent state-of-the-art neural text matching models (PLMs) are hard to generalize to different tasks.
We adopt a specialization-generalization training strategy and refer to it as Match-Prompt.
In specialization stage, descriptions of different matching tasks are mapped to only a few prompt tokens.
In generalization stage, text matching model explores the essential matching signals by being trained on diverse multiple matching tasks.
arXiv Detail & Related papers (2022-04-06T11:01:08Z) - Simple Contrastive Representation Adversarial Learning for NLP Tasks [17.12062566060011]
Two novel frameworks, supervised contrastive adversarial learning (SCAL) and unsupervised SCAL (USCAL), are proposed.
We employ it to Transformer-based models for natural language understanding, sentence semantic textual similarity and adversarial learning tasks.
Experimental results on GLUE benchmark tasks show that our fine-tuned supervised method outperforms BERT$_base$ over 1.75%.
arXiv Detail & Related papers (2021-11-26T03:16:09Z) - Distribution Matching for Heterogeneous Multi-Task Learning: a
Large-scale Face Study [75.42182503265056]
Multi-Task Learning has emerged as a methodology in which multiple tasks are jointly learned by a shared learning algorithm.
We deal with heterogeneous MTL, simultaneously addressing detection, classification & regression problems.
We build FaceBehaviorNet, the first framework for large-scale face analysis, by jointly learning all facial behavior tasks.
arXiv Detail & Related papers (2021-05-08T22:26:52Z) - ERICA: Improving Entity and Relation Understanding for Pre-trained
Language Models via Contrastive Learning [97.10875695679499]
We propose a novel contrastive learning framework named ERICA in pre-training phase to obtain a deeper understanding of the entities and their relations in text.
Experimental results demonstrate that our proposed ERICA framework achieves consistent improvements on several document-level language understanding tasks.
arXiv Detail & Related papers (2020-12-30T03:35:22Z) - Learning not to Discriminate: Task Agnostic Learning for Improving
Monolingual and Code-switched Speech Recognition [12.354292498112347]
We present further improvements over our previous work by using domain adversarial learning to train task models.
Our proposed technique leads to reductions in Word Error Rates (WER) in monolingual and code-switched test sets across three language pairs.
arXiv Detail & Related papers (2020-06-09T13:45:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.