YNU-HPCC at SemEval-2020 Task 11: LSTM Network for Detection of
Propaganda Techniques in News Articles
- URL: http://arxiv.org/abs/2008.10166v2
- Date: Tue, 25 Aug 2020 11:03:18 GMT
- Title: YNU-HPCC at SemEval-2020 Task 11: LSTM Network for Detection of
Propaganda Techniques in News Articles
- Authors: Jiaxu Dao, Jin Wang, Xuejie Zhang
- Abstract summary: This paper summarizes our studies on propaganda detection techniques for news articles in the SemEval-2020 task 11.
We implement the GloVe word representation, the BERT pretraining model, and the LSTM model architecture to accomplish this task.
Our method significantly outperforms the officially released baseline method, and the SI and TC subtasks rank 17th and 22nd, respectively, for the test set.
- Score: 5.352512345142247
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper summarizes our studies on propaganda detection techniques for news
articles in the SemEval-2020 task 11. This task is divided into the SI and TC
subtasks. We implemented the GloVe word representation, the BERT pretraining
model, and the LSTM model architecture to accomplish this task. Our approach
achieved good results for both the SI and TC subtasks. The macro-F1-score for
the SI subtask is 0.406, and the micro-F1-score for the TC subtask is 0.505.
Our method significantly outperforms the officially released baseline method,
and the SI and TC subtasks rank 17th and 22nd, respectively, for the test set.
This paper also compares the performances of different deep learning model
architectures, such as the Bi-LSTM, LSTM, BERT, and XGBoost models, on the
detection of news promotion techniques. The code of this paper is availabled
at: https://github.com/daojiaxu/semeval_11.
Related papers
- RoBLEURT Submission for the WMT2021 Metrics Task [72.26898579202076]
We present our submission to the Shared Metrics Task: RoBLEURT.
Our model reaches state-of-the-art correlations with the WMT 2020 human annotations upon 8 out of 10 to-English language pairs.
arXiv Detail & Related papers (2022-04-28T08:49:40Z) - Learning to Win Lottery Tickets in BERT Transfer via Task-agnostic Mask
Training [55.43088293183165]
Recent studies show that pre-trained language models (PLMs) like BERT contain matchingworks that have similar transfer learning performance as the original PLM.
In this paper, we find that the BERTworks have even more potential than these studies have shown.
We train binary masks over model weights on the pre-training tasks, with the aim of preserving the universal transferability of the subnetwork.
arXiv Detail & Related papers (2022-04-24T08:42:47Z) - Improving Deep Learning for HAR with shallow LSTMs [70.94062293989832]
We propose to alter the DeepConvLSTM to employ a 1-layered instead of a 2-layered LSTM.
Our results stand in contrast to the belief that one needs at least a 2-layered LSTM when dealing with sequential data.
arXiv Detail & Related papers (2021-08-02T08:14:59Z) - The USYD-JD Speech Translation System for IWSLT 2021 [85.64797317290349]
This paper describes the University of Sydney& JD's joint submission of the IWSLT 2021 low resource speech translation task.
We trained our models with the officially provided ASR and MT datasets.
To achieve better translation performance, we explored the most recent effective strategies, including back translation, knowledge distillation, multi-feature reranking and transductive finetuning.
arXiv Detail & Related papers (2021-07-24T09:53:34Z) - TechTexC: Classification of Technical Texts using Convolution and
Bidirectional Long Short Term Memory Network [0.0]
A classification system (called 'TechTexC') is developed to perform the classification task using three techniques.
Results show that CNN with BiLSTM model outperforms the other techniques concerning task-1 of sub-tasks (a, b, c and g) and task-2a.
In the case of test set, the combined CNN with BiLSTM approach achieved that higher accuracy for the subtasks 1a (70.76%), 1b (79.97%), 1c (65.45%), 1g (49.23%) and 2a (70.14%)
arXiv Detail & Related papers (2020-12-21T15:22:47Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z) - syrapropa at SemEval-2020 Task 11: BERT-based Models Design For
Propagandistic Technique and Span Detection [2.0051855303186046]
We first build the model for Span Identification (SI) based on SpanBERT, and facilitate the detection by a deeper model and a sentence-level representation.
We then develop a hybrid model for the Technique Classification (TC)
The hybrid model is composed of three submodels including two BERT models with different training methods, and a feature-based Logistic Regression model.
arXiv Detail & Related papers (2020-08-24T02:15:29Z) - DUTH at SemEval-2020 Task 11: BERT with Entity Mapping for Propaganda
Classification [1.5469452301122173]
This report describes the methods employed by the Democritus University of Thrace (DUTH) team for participating in SemEval-2020 Task 11: Detection of Propaganda Techniques in News Articles.
arXiv Detail & Related papers (2020-08-22T18:18:02Z) - CyberWallE at SemEval-2020 Task 11: An Analysis of Feature Engineering
for Ensemble Models for Propaganda Detection [0.0]
We use a bi-LSTM architecture in the Span Identification subtask and train a complex ensemble model for the Technique Classification subtask.
Our systems achieve a rank of 8 out of 35 teams in the SI subtask and 8 out of 31 teams in the TC subtask.
arXiv Detail & Related papers (2020-08-22T15:51:16Z) - NoPropaganda at SemEval-2020 Task 11: A Borrowed Approach to Sequence
Tagging and Text Classification [0.0]
This paper describes our contribution to SemEval-2020 Task 11: Detection Of Propaganda Techniques In News Articles.
We start with simple LSTM baselines and move to an autoregressive transformer decoder to predict long continuous propaganda spans for the first subtask.
We also adopt an approach from relation extraction by enveloping spans mentioned above with special tokens for the second subtask of propaganda technique classification.
arXiv Detail & Related papers (2020-07-25T11:35:57Z) - Device-Robust Acoustic Scene Classification Based on Two-Stage
Categorization and Data Augmentation [63.98724740606457]
We present a joint effort of four groups, namely GT, USTC, Tencent, and UKE, to tackle Task 1 - Acoustic Scene Classification (ASC) in the DCASE 2020 Challenge.
Task 1a focuses on ASC of audio signals recorded with multiple (real and simulated) devices into ten different fine-grained classes.
Task 1b concerns with classification of data into three higher-level classes using low-complexity solutions.
arXiv Detail & Related papers (2020-07-16T15:07:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.