Few-shot learning for automated content analysis: Efficient coding of
arguments and claims in the debate on arms deliveries to Ukraine
- URL: http://arxiv.org/abs/2312.16975v1
- Date: Thu, 28 Dec 2023 11:39:08 GMT
- Title: Few-shot learning for automated content analysis: Efficient coding of
arguments and claims in the debate on arms deliveries to Ukraine
- Authors: Jonas Rieger, Kostiantyn Yanchenko, Mattes Ruckdeschel, Gerret von
Nordheim, Katharina Kleinen-von K\"onigsl\"ow, Gregor Wiedemann
- Abstract summary: Pre-trained language models (PLM) based on transformer neural networks offer great opportunities to improve automatic content analysis in communication science.
Three characteristics so far impeded the widespread adoption of the methods in the applying disciplines: the dominance of English language models in NLP research, the necessary computing resources, and the effort required to produce training data to fine-tune PLMs.
We test our approach on a realistic use case from communication science to automatically detect claims and arguments together with their stance in the German news debate on arms deliveries to Ukraine.
- Score: 0.9576975587953563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pre-trained language models (PLM) based on transformer neural networks
developed in the field of natural language processing (NLP) offer great
opportunities to improve automatic content analysis in communication science,
especially for the coding of complex semantic categories in large datasets via
supervised machine learning. However, three characteristics so far impeded the
widespread adoption of the methods in the applying disciplines: the dominance
of English language models in NLP research, the necessary computing resources,
and the effort required to produce training data to fine-tune PLMs. In this
study, we address these challenges by using a multilingual transformer model in
combination with the adapter extension to transformers, and few-shot learning
methods. We test our approach on a realistic use case from communication
science to automatically detect claims and arguments together with their stance
in the German news debate on arms deliveries to Ukraine. In three experiments,
we evaluate (1) data preprocessing strategies and model variants for this task,
(2) the performance of different few-shot learning methods, and (3) how well
the best setup performs on varying training set sizes in terms of validity,
reliability, replicability and reproducibility of the results. We find that our
proposed combination of transformer adapters with pattern exploiting training
provides a parameter-efficient and easily shareable alternative to fully
fine-tuning PLMs. It performs on par in terms of validity, while overall,
provides better properties for application in communication studies. The
results also show that pre-fine-tuning for a task on a near-domain dataset
leads to substantial improvement, in particular in the few-shot setting.
Further, the results indicate that it is useful to bias the dataset away from
the viewpoints of specific prominent individuals.
Related papers
- Rephrase and Contrast: Fine-Tuning Language Models for Enhanced Understanding of Communication and Computer Networks [13.829525575305206]
This paper introduces our Rephrase and Contrast (RaC) framework, an efficient fine-tuning framework.
RaC enhances LLMs' comprehension and critical thinking abilities by incorporating question reformulation and contrastive analysis.
To efficiently construct the dataset for RaC fine-tuning, we develop a GPT-assisted data mining method for generating high-quality question-answer pairs.
arXiv Detail & Related papers (2024-09-21T16:04:43Z) - Supervised Pretraining Can Learn In-Context Reinforcement Learning [96.62869749926415]
In this paper, we study the in-context learning capabilities of transformers in decision-making problems.
We introduce and study Decision-Pretrained Transformer (DPT), a supervised pretraining method where the transformer predicts an optimal action.
We find that the pretrained transformer can be used to solve a range of RL problems in-context, exhibiting both exploration online and conservatism offline.
arXiv Detail & Related papers (2023-06-26T17:58:50Z) - Pre-trained Language Models for Keyphrase Generation: A Thorough
Empirical Study [76.52997424694767]
We present an in-depth empirical study of keyphrase extraction and keyphrase generation using pre-trained language models.
We show that PLMs have competitive high-resource performance and state-of-the-art low-resource performance.
Further results show that in-domain BERT-like PLMs can be used to build strong and data-efficient keyphrase generation models.
arXiv Detail & Related papers (2022-12-20T13:20:21Z) - On the Usability of Transformers-based models for a French
Question-Answering task [2.44288434255221]
This paper focuses on the usability of Transformer-based language models in small-scale learning problems.
We introduce a new compact model for French FrALBERT which proves to be competitive in low-resource settings.
arXiv Detail & Related papers (2022-07-19T09:46:15Z) - Improving Classifier Training Efficiency for Automatic Cyberbullying
Detection with Feature Density [58.64907136562178]
We study the effectiveness of Feature Density (FD) using different linguistically-backed feature preprocessing methods.
We hypothesise that estimating dataset complexity allows for the reduction of the number of required experiments.
The difference in linguistic complexity of datasets allows us to additionally discuss the efficacy of linguistically-backed word preprocessing.
arXiv Detail & Related papers (2021-11-02T15:48:28Z) - Efficient Nearest Neighbor Language Models [114.40866461741795]
Non-parametric neural language models (NLMs) learn predictive distributions of text utilizing an external datastore.
We show how to achieve up to a 6x speed-up in inference speed while retaining comparable performance.
arXiv Detail & Related papers (2021-09-09T12:32:28Z) - SDA: Improving Text Generation with Self Data Augmentation [88.24594090105899]
We propose to improve the standard maximum likelihood estimation (MLE) paradigm by incorporating a self-imitation-learning phase for automatic data augmentation.
Unlike most existing sentence-level augmentation strategies, our method is more general and could be easily adapted to any MLE-based training procedure.
arXiv Detail & Related papers (2021-01-02T01:15:57Z) - Fine-tuning BERT for Low-Resource Natural Language Understanding via
Active Learning [30.5853328612593]
In this work, we explore fine-tuning methods of BERT -- a pre-trained Transformer based language model.
Our experimental results show an advantage in model performance by maximizing the approximate knowledge gain of the model.
We analyze the benefits of freezing layers of the language model during fine-tuning to reduce the number of trainable parameters.
arXiv Detail & Related papers (2020-12-04T08:34:39Z) - Dynamic Sampling and Selective Masking for Communication-Efficient
Federated Learning [11.511755449420253]
Federated learning (FL) is a novel machine learning setting that enables on-device intelligence via decentralized training and federated optimization.
This paper introduces two approaches for improving communication efficiency by dynamic sampling and top-$k$ selective masking.
arXiv Detail & Related papers (2020-03-21T08:20:04Z) - Hybrid Generative-Retrieval Transformers for Dialogue Domain Adaptation [77.62366712130196]
We present the winning entry at the fast domain adaptation task of DSTC8, a hybrid generative-retrieval model based on GPT-2 fine-tuned to the multi-domain MetaLWOz dataset.
Our model uses retrieval logic as a fallback, being SoTA on MetaLWOz in human evaluation (>4% improvement over the 2nd place system) and attaining competitive generalization performance in adaptation to the unseen MultiWOZ dataset.
arXiv Detail & Related papers (2020-03-03T18:07:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.