DUTH at SemEval-2020 Task 11: BERT with Entity Mapping for Propaganda
Classification
- URL: http://arxiv.org/abs/2008.09894v2
- Date: Tue, 25 Aug 2020 10:20:22 GMT
- Title: DUTH at SemEval-2020 Task 11: BERT with Entity Mapping for Propaganda
Classification
- Authors: Anastasios Bairaktaris, Symeon Symeonidis, Avi Arampatzis
- Abstract summary: This report describes the methods employed by the Democritus University of Thrace (DUTH) team for participating in SemEval-2020 Task 11: Detection of Propaganda Techniques in News Articles.
- Score: 1.5469452301122173
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This report describes the methods employed by the Democritus University of
Thrace (DUTH) team for participating in SemEval-2020 Task 11: Detection of
Propaganda Techniques in News Articles. Our team dealt with Subtask 2:
Technique Classification. We used shallow Natural Language Processing (NLP)
preprocessing techniques to reduce the noise in the dataset, feature selection
methods, and common supervised machine learning algorithms. Our final model is
based on using the BERT system with entity mapping. To improve our model's
accuracy, we mapped certain words into five distinct categories by employing
word-classes and entity recognition.
Related papers
- Automated Few-shot Classification with Instruction-Finetuned Language
Models [76.69064714392165]
We show that AuT-Few outperforms state-of-the-art few-shot learning methods.
We also show that AuT-Few is the best ranking method across datasets on the RAFT few-shot benchmark.
arXiv Detail & Related papers (2023-05-21T21:50:27Z) - Neural Coreference Resolution based on Reinforcement Learning [53.73316523766183]
Coreference resolution systems need to solve two subtasks.
One task is to detect all of the potential mentions, the other is to learn the linking of an antecedent for each possible mention.
We propose a reinforcement learning actor-critic-based neural coreference resolution system.
arXiv Detail & Related papers (2022-12-18T07:36:35Z) - Training ELECTRA Augmented with Multi-word Selection [53.77046731238381]
We present a new text encoder pre-training method that improves ELECTRA based on multi-task learning.
Specifically, we train the discriminator to simultaneously detect replaced tokens and select original tokens from candidate sets.
arXiv Detail & Related papers (2021-05-31T23:19:00Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - Solomon at SemEval-2020 Task 11: Ensemble Architecture for Fine-Tuned
Propaganda Detection in News Articles [0.3232625980782302]
This paper describes our system (Solomon) details and results of participation in the SemEval 2020 Task 11 "Detection of Propaganda Techniques in News Articles"
We used RoBERTa based transformer architecture for fine-tuning on the propaganda dataset.
Compared to the other participating systems, our submission is ranked 4th on the leaderboard.
arXiv Detail & Related papers (2020-09-16T05:00:40Z) - Team DoNotDistribute at SemEval-2020 Task 11: Features, Finetuning, and
Data Augmentation in Neural Models for Propaganda Detection in News Articles [13.339333273943843]
This paper presents our systems for SemEval 2020 Shared Task 11: Detection of Propaganda Techniques in News Articles.
We participate in both the span identification and technique classification subtasks and report on experiments using different BERT-based models along with handcrafted features.
arXiv Detail & Related papers (2020-08-21T22:35:57Z) - LTIatCMU at SemEval-2020 Task 11: Incorporating Multi-Level Features for
Multi-Granular Propaganda Span Identification [70.1903083747775]
This paper describes our submission for the task of Propaganda Span Identification in news articles.
We introduce a BERT-BiLSTM based span-level propaganda classification model that identifies which token spans within the sentence are indicative of propaganda.
arXiv Detail & Related papers (2020-08-11T16:14:47Z) - NoPropaganda at SemEval-2020 Task 11: A Borrowed Approach to Sequence
Tagging and Text Classification [0.0]
This paper describes our contribution to SemEval-2020 Task 11: Detection Of Propaganda Techniques In News Articles.
We start with simple LSTM baselines and move to an autoregressive transformer decoder to predict long continuous propaganda spans for the first subtask.
We also adopt an approach from relation extraction by enveloping spans mentioned above with special tokens for the second subtask of propaganda technique classification.
arXiv Detail & Related papers (2020-07-25T11:35:57Z) - newsSweeper at SemEval-2020 Task 11: Context-Aware Rich Feature
Representations For Propaganda Classification [2.0491741153610334]
This paper describes our submissions to SemEval 2020 Task 11: Detection of Propaganda Techniques in News Articles.
We make use of pre-trained BERT language model enhanced with tagging techniques developed for the task of Named Entity Recognition.
For the second subtask, we incorporate contextual features in a pre-trained RoBERTa model for the classification of propaganda techniques.
arXiv Detail & Related papers (2020-07-21T14:06:59Z) - BPGC at SemEval-2020 Task 11: Propaganda Detection in News Articles with
Multi-Granularity Knowledge Sharing and Linguistic Features based Ensemble
Learning [2.8913142991383114]
SemEval 2020 Task-11 aims to design automated systems for news propaganda detection.
Task-11 consists of two sub-tasks, namely, Span Identification and Technique Classification.
arXiv Detail & Related papers (2020-05-31T19:35:53Z) - Kungfupanda at SemEval-2020 Task 12: BERT-Based Multi-Task Learning for
Offensive Language Detection [55.445023584632175]
We build an offensive language detection system, which combines multi-task learning with BERT-based models.
Our model achieves 91.51% F1 score in English Sub-task A, which is comparable to the first place.
arXiv Detail & Related papers (2020-04-28T11:27:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.