Learning for Detecting Norm Violation in Online Communities
- URL: http://arxiv.org/abs/2104.14911v1
- Date: Fri, 30 Apr 2021 11:18:04 GMT
- Title: Learning for Detecting Norm Violation in Online Communities
- Authors: Thiago Freitas dos Santos, Nardine Osman and Marco Schorlemmer
- Abstract summary: We propose a framework capable of detecting norm violations in online communities.
We build our framework using Machine Learning, with Logistic Model Trees as the classification algorithm.
Our work is then evaluated with the Wikipedia use case where we focus on the norm that prohibits vandalism in Wikipedia edits.
- Score: 3.480626767752489
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we focus on normative systems for online communities. The
paper addresses the issue that arises when different community members
interpret these norms in different ways, possibly leading to unexpected
behavior in interactions, usually with norm violations that affect the
individual and community experiences. To address this issue, we propose a
framework capable of detecting norm violations and providing the violator with
information about the features of their action that makes this action violate a
norm. We build our framework using Machine Learning, with Logistic Model Trees
as the classification algorithm. Since norm violations can be highly
contextual, we train our model using data from the Wikipedia online community,
namely data on Wikipedia edits. Our work is then evaluated with the Wikipedia
use case where we focus on the norm that prohibits vandalism in Wikipedia
edits.
Related papers
- Learning Rules from KGs Guided by Language Models [48.858741745144044]
Rule learning methods can be applied to predict potentially missing facts.
Ranking of rules is especially challenging over highly incomplete or biased KGs.
With the recent rise of Language Models (LMs) several works have claimed that LMs can be used as alternative means for KG completion.
arXiv Detail & Related papers (2024-09-12T09:27:36Z) - HelloFresh: LLM Evaluations on Streams of Real-World Human Editorial Actions across X Community Notes and Wikipedia edits [92.62157408704594]
HelloFresh is based on continuous streams of real-world data generated by intrinsically motivated human labelers.
It covers recent events from X (formerly Twitter) community notes and edits of Wikipedia pages.
It mitigates the risk of test data contamination and benchmark overfitting.
arXiv Detail & Related papers (2024-06-05T16:25:57Z) - SADAS: A Dialogue Assistant System Towards Remediating Norm Violations
in Bilingual Socio-Cultural Conversations [56.31816995795216]
Socially-Aware Dialogue Assistant System (SADAS) is designed to ensure that conversations unfold with respect and understanding.
Our system's novel architecture includes: (1) identifying the categories of norms present in the dialogue, (2) detecting potential norm violations, (3) evaluating the severity of these violations, and (4) implementing targeted remedies to rectify the breaches.
arXiv Detail & Related papers (2024-01-29T08:54:21Z) - Reading Books is Great, But Not if You Are Driving! Visually Grounded
Reasoning about Defeasible Commonsense Norms [65.17491295329991]
We construct a new benchmark for studying visual-grounded commonsense norms: NORMLENS.
We find that state-of-the-art model judgments and explanations are not well-aligned with human annotation.
We present a new approach to better align models with humans by distilling social commonsense knowledge from large language models.
arXiv Detail & Related papers (2023-10-16T14:00:07Z) - Analyzing Norm Violations in Live-Stream Chat [49.120561596550395]
We study the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms.
We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch.
Our results show that appropriate contextual information can boost moderation performance by 35%.
arXiv Detail & Related papers (2023-05-18T05:58:27Z) - CPL-NoViD: Context-Aware Prompt-based Learning for Norm Violation Detection in Online Communities [28.576099654579437]
We introduce Context-aware Prompt-based Learning for Norm Violation Detection (CPL-NoViD)
CPL-NoViD outperforms the baseline by incorporating context through natural language prompts.
It establishes a new state-of-the-art in norm violation detection, surpassing existing benchmarks.
arXiv Detail & Related papers (2023-05-16T23:27:59Z) - Explainable Abuse Detection as Intent Classification and Slot Filling [66.80201541759409]
We introduce the concept of policy-aware abuse detection, abandoning the unrealistic expectation that systems can reliably learn which phenomena constitute abuse from inspecting the data alone.
We show how architectures for intent classification and slot filling can be used for abuse detection, while providing a rationale for model decisions.
arXiv Detail & Related papers (2022-10-06T03:33:30Z) - Detecting Community Sensitive Norm Violations in Online Conversations [21.892867827127603]
We focus on a more complete spectrum of community norms and their violations in the local conversational and global community contexts.
We introduce a series of models that use this data to develop context- and community-sensitive norm violation detection.
arXiv Detail & Related papers (2021-10-09T00:39:35Z) - On Analyzing Annotation Consistency in Online Abusive Behavior Datasets [5.900114841365645]
Researchers have proposed, collected, and annotated online abusive content datasets.
These datasets play a critical role in facilitating the research on online hate speech and abusive behaviors.
It is often contentious on what should be the true label of a given text as the semantic difference of the labels may be blurred.
arXiv Detail & Related papers (2020-06-24T06:34:25Z) - WAC: A Corpus of Wikipedia Conversations for Online Abuse Detection [0.0]
We propose an original framework, based on the Wikipedia Comment corpus, with comment-level annotations of different types.
This large corpus of more than 380k annotated messages opens perspectives for online abuse detection and especially for context-based approaches.
We also propose, in addition to this corpus, a complete benchmarking platform to stimulate and fairly compare scientific works around the problem of content abuse detection.
arXiv Detail & Related papers (2020-03-13T10:26:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.