The Topic Confusion Task: A Novel Scenario for Authorship Attribution
- URL: http://arxiv.org/abs/2104.08530v1
- Date: Sat, 17 Apr 2021 12:50:58 GMT
- Title: The Topic Confusion Task: A Novel Scenario for Authorship Attribution
- Authors: Malik H. Altakrori (1 and 3), Jackie Chi Kit Cheung (1 and 3),
Benjamin C. M. Fung (2 and 3) ((1) School of Computer Science -McGill
University, (2) School of Information Studies-McGill University, (3) Mila)
- Abstract summary: Authorship attribution is the problem of identifying the most plausible author of an anonymous text from a set of candidate authors.
We propose the emphtopic confusion task, where we switch the author-topic configuration between training and testing set.
By evaluating different features, we show that stylometric features with part-of-speech tags are less susceptible to topic variations and can increase the accuracy of the attribution process.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Authorship attribution is the problem of identifying the most plausible
author of an anonymous text from a set of candidate authors. Researchers have
investigated same-topic and cross-topic scenarios of authorship attribution,
which differ according to whether unseen topics are used in the testing phase.
However, neither scenario allows us to explain whether errors are caused by
failure to capture authorship style, by the topic shift or by other factors.
Motivated by this, we propose the \emph{topic confusion} task, where we switch
the author-topic configuration between training and testing set. This setup
allows us to probe errors in the attribution process. We investigate the
accuracy and two error measures: one caused by the models' confusion by the
switch because the features capture the topics, and one caused by the features'
inability to capture the writing styles, leading to weaker models. By
evaluating different features, we show that stylometric features with
part-of-speech tags are less susceptible to topic variations and can increase
the accuracy of the attribution process. We further show that combining them
with word-level $n$-grams can outperform the state-of-the-art technique in the
cross-topic scenario. Finally, we show that pretrained language models such as
BERT and RoBERTa perform poorly on this task, and are outperformed by simple
$n$-gram features.
Related papers
- Attacking Misinformation Detection Using Adversarial Examples Generated by Language Models [0.0]
We investigate the challenge of generating adversarial examples to test the robustness of text classification algorithms.
We focus on simulation of content moderation by setting realistic limits on the number of queries an attacker is allowed to attempt.
arXiv Detail & Related papers (2024-10-28T11:46:30Z) - IDT: Dual-Task Adversarial Attacks for Privacy Protection [8.312362092693377]
Methods to protect privacy can involve using representations inside models that are not to detect sensitive attributes.
We propose IDT, a method that analyses predictions made by auxiliary and interpretable models to identify which tokens are important to change.
We evaluate different datasets for NLP suitable for different tasks.
arXiv Detail & Related papers (2024-06-28T04:14:35Z) - Forging the Forger: An Attempt to Improve Authorship Verification via Data Augmentation [52.72682366640554]
Authorship Verification (AV) is a text classification task concerned with inferring whether a candidate text has been written by one specific author or by someone else.
It has been shown that many AV systems are vulnerable to adversarial attacks, where a malicious author actively tries to fool the classifier by either concealing their writing style, or by imitating the style of another author.
arXiv Detail & Related papers (2024-03-17T16:36:26Z) - Can Authorship Attribution Models Distinguish Speakers in Speech Transcripts? [4.148732457277201]
Authorship verification is the task of determining if two distinct writing samples share the same author.
In this paper, we explore the attribution of transcribed speech, which poses novel challenges.
We propose a new benchmark for speaker attribution focused on human-transcribed conversational speech transcripts.
arXiv Detail & Related papers (2023-11-13T18:54:17Z) - PRIME: Prioritizing Interpretability in Failure Mode Extraction [49.93565079216376]
We study the challenge of providing human-understandable descriptions for failure modes in trained image classification models.
We propose a novel approach that prioritizes interpretability in this problem.
Our method successfully identifies failure modes and generates high-quality text descriptions associated with them.
arXiv Detail & Related papers (2023-09-29T22:00:12Z) - Verifying the Robustness of Automatic Credibility Assessment [79.08422736721764]
Text classification methods have been widely investigated as a way to detect content of low credibility.
In some cases insignificant changes in input text can mislead the models.
We introduce BODEGA: a benchmark for testing both victim models and attack methods on misinformation detection tasks.
arXiv Detail & Related papers (2023-03-14T16:11:47Z) - PART: Pre-trained Authorship Representation Transformer [64.78260098263489]
Authors writing documents imprint identifying information within their texts: vocabulary, registry, punctuation, misspellings, or even emoji usage.
Previous works use hand-crafted features or classification tasks to train their authorship models, leading to poor performance on out-of-domain authors.
We propose a contrastively trained model fit to learn textbfauthorship embeddings instead of semantics.
arXiv Detail & Related papers (2022-09-30T11:08:39Z) - Experiments with adversarial attacks on text genres [0.0]
Neural models based on pre-trained transformers, such as BERT or XLM-RoBERTa, demonstrate SOTA results in many NLP tasks.
We show that embedding-based algorithms which can replace some of the most significant'' words with words similar to them, have the ability to influence model predictions in a significant proportion of cases.
arXiv Detail & Related papers (2021-07-05T19:37:59Z) - Towards Variable-Length Textual Adversarial Attacks [68.27995111870712]
It is non-trivial to conduct textual adversarial attacks on natural language processing tasks due to the discreteness of data.
In this paper, we propose variable-length textual adversarial attacks(VL-Attack)
Our method can achieve $33.18$ BLEU score on IWSLT14 German-English translation, achieving an improvement of $1.47$ over the baseline model.
arXiv Detail & Related papers (2021-04-16T14:37:27Z) - Narrative Incoherence Detection [76.43894977558811]
We propose the task of narrative incoherence detection as a new arena for inter-sentential semantic understanding.
Given a multi-sentence narrative, decide whether there exist any semantic discrepancies in the narrative flow.
arXiv Detail & Related papers (2020-12-21T07:18:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.