Harmonizing Global Voices: Culturally-Aware Models for Enhanced Content
Moderation
- URL: http://arxiv.org/abs/2312.02401v1
- Date: Tue, 5 Dec 2023 00:11:09 GMT
- Title: Harmonizing Global Voices: Culturally-Aware Models for Enhanced Content
Moderation
- Authors: Alex J. Chan, Jos\'e Luis Redondo Garc\'ia, Fabrizio Silvestri, Colm
O'Donnel, Konstantina Palla
- Abstract summary: We train large language models on extensive datasets of media news and articles to create culturally attuned models.
We find that training on extensive media datasets successfully induced cultural awareness and resulted in improvements in handling content violations on a regional basis.
- Score: 10.53562175155486
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Content moderation at scale faces the challenge of considering local cultural
distinctions when assessing content. While global policies aim to maintain
decision-making consistency and prevent arbitrary rule enforcement, they often
overlook regional variations in interpreting natural language as expressed in
content. In this study, we are looking into how moderation systems can tackle
this issue by adapting to local comprehension nuances. We train large language
models on extensive datasets of media news and articles to create culturally
attuned models. The latter aim to capture the nuances of communication across
geographies with the goal of recognizing cultural and societal variations in
what is considered offensive content. We further explore the capability of
these models to generate explanations for instances of content violation,
aiming to shed light on how policy guidelines are perceived when cultural and
societal contexts change. We find that training on extensive media datasets
successfully induced cultural awareness and resulted in improvements in
handling content violations on a regional basis. Additionally, these
advancements include the ability to provide explanations that align with the
specific local norms and nuances as evidenced by the annotators' preference in
our conducted study. This multifaceted success reinforces the critical role of
an adaptable content moderation approach in keeping pace with the ever-evolving
nature of the content it oversees.
Related papers
- Methodology of Adapting Large English Language Models for Specific Cultural Contexts [10.151487049108626]
We propose a rapid adaptation method for large models in specific cultural contexts.
The adapted LLM significantly enhances its capabilities in domain-specific knowledge and adaptability to safety values.
arXiv Detail & Related papers (2024-06-26T09:16:08Z) - Extrinsic Evaluation of Cultural Competence in Large Language Models [53.626808086522985]
We focus on extrinsic evaluation of cultural competence in two text generation tasks.
We evaluate model outputs when an explicit cue of culture, specifically nationality, is perturbed in the prompts.
We find weak correlations between text similarity of outputs for different countries and the cultural values of these countries.
arXiv Detail & Related papers (2024-06-17T14:03:27Z) - CIVICS: Building a Dataset for Examining Culturally-Informed Values in Large Language Models [59.22460740026037]
"CIVICS: Culturally-Informed & Values-Inclusive Corpus for Societal impacts" dataset is designed to evaluate the social and cultural variation of Large Language Models (LLMs)
We create a hand-crafted, multilingual dataset of value-laden prompts which address specific socially sensitive topics, including LGBTQI rights, social welfare, immigration, disability rights, and surrogacy.
arXiv Detail & Related papers (2024-05-22T20:19:10Z) - The Unappreciated Role of Intent in Algorithmic Moderation of Social Media Content [2.2618341648062477]
This paper examines the role of intent in content moderation systems.
We review state of the art detection models and benchmark training datasets for online abuse to assess their awareness and ability to capture intent.
arXiv Detail & Related papers (2024-05-17T18:05:13Z) - CRAFT: Extracting and Tuning Cultural Instructions from the Wild [38.255242754975654]
This paper introduces a novel pipeline for extracting high-quality, culturally-related instruction tuning datasets from vast unstructured corpora.
We utilize a self-instruction generation pipeline to identify cultural concepts and trigger instruction.
We conduct experiments across three regions: Singapore, the Philippines, and the United States, achieving performance improvement of up to 6%.
arXiv Detail & Related papers (2024-05-06T03:21:55Z) - Massively Multi-Cultural Knowledge Acquisition & LM Benchmarking [48.21982147529661]
This paper introduces a novel approach for massively multicultural knowledge acquisition.
Our method strategically navigates from densely informative Wikipedia documents on cultural topics to an extensive network of linked pages.
Our work marks an important step towards deeper understanding and bridging the gaps of cultural disparities in AI.
arXiv Detail & Related papers (2024-02-14T18:16:54Z) - Not All Countries Celebrate Thanksgiving: On the Cultural Dominance in
Large Language Models [89.94270049334479]
This paper identifies a cultural dominance issue within large language models (LLMs)
LLMs often provide inappropriate English-culture-related answers that are not relevant to the expected culture when users ask in non-English languages.
arXiv Detail & Related papers (2023-10-19T05:38:23Z) - Natural Language Decompositions of Implicit Content Enable Better Text
Representations [56.85319224208865]
We introduce a method for the analysis of text that takes implicitly communicated content explicitly into account.
We use a large language model to produce sets of propositions that are inferentially related to the text that has been observed.
Our results suggest that modeling the meanings behind observed language, rather than the literal text alone, is a valuable direction for NLP.
arXiv Detail & Related papers (2023-05-23T23:45:20Z) - Countering Malicious Content Moderation Evasion in Online Social
Networks: Simulation and Detection of Word Camouflage [64.78260098263489]
Twisting and camouflaging keywords are among the most used techniques to evade platform content moderation systems.
This article contributes significantly to countering malicious information by developing multilingual tools to simulate and detect new methods of evasion of content.
arXiv Detail & Related papers (2022-12-27T16:08:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.