Anti-Sexism Alert System: Identification of Sexist Comments on Social
Media Using AI Techniques
- URL: http://arxiv.org/abs/2312.00053v1
- Date: Tue, 28 Nov 2023 19:48:46 GMT
- Title: Anti-Sexism Alert System: Identification of Sexist Comments on Social
Media Using AI Techniques
- Authors: Rebeca P. D\'iaz Redondo and Ana Fern\'andez Vilas and Mateo Ramos
Merino and Sonia Valladares and Soledad Torres Guijarro and Manar Mohamed
Hafez
- Abstract summary: Sexist comments that are publicly posted in social media (newspaper comments, social networks, etc.) usually obtain a lot of attention and become viral, with consequent damage to the persons involved.
In this paper, we introduce an anti-sexism alert system, based on natural language processing (NLP) and artificial intelligence (AI)
This system analyzes any public post, and decides if it could be considered a sexist comment or not.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Social relationships in the digital sphere are becoming more usual and
frequent, and they constitute a very important aspect for all of us. {Violent
interactions in this sphere are very frequent, and have serious effects on the
victims}. Within this global scenario, there is one kind of digital violence
that is becoming really worrying: sexism against women. Sexist comments that
are publicly posted in social media (newspaper comments, social networks,
etc.), usually obtain a lot of attention and become viral, with consequent
damage to the persons involved. In this paper, we introduce an anti-sexism
alert system, based on natural language processing (NLP) and artificial
intelligence (AI), that analyzes any public post, and decides if it could be
considered a sexist comment or not. Additionally, this system also works on
analyzing all the public comments linked to any multimedia content (piece of
news, video, tweet, etc.) and decides, using a color-based system similar to
traffic lights, if there is sexism in the global set of posts. We have created
a labeled data set in Spanish, since the majority of studies focus on English,
to train our system, which offers a very good performance after the validation
experiments.
Related papers
- A Holistic Indicator of Polarization to Measure Online Sexism [2.498836880652668]
The online trend of the manosphere and feminist discourse on social networks requires a holistic measure of the level of sexism in an online community.
This indicator is important for policymakers and moderators of online communities.
We build a model that can provide a comparable holistic indicator of toxicity targeted toward male and female identity and male and female individuals.
arXiv Detail & Related papers (2024-04-02T18:00:42Z) - Exploratory Data Analysis on Code-mixed Misogynistic Comments [0.0]
We present a novel dataset of YouTube comments in mix-code Hinglish.
These comments have been weak labelled as Misogynistic' and Non-misogynistic'
arXiv Detail & Related papers (2024-03-09T23:21:17Z) - "I'm fully who I am": Towards Centering Transgender and Non-Binary
Voices to Measure Biases in Open Language Generation [69.25368160338043]
Transgender and non-binary (TGNB) individuals disproportionately experience discrimination and exclusion from daily life.
We assess how the social reality surrounding experienced marginalization of TGNB persons contributes to and persists within Open Language Generation.
We introduce TANGO, a dataset of template-based real-world text curated from a TGNB-oriented community.
arXiv Detail & Related papers (2023-05-17T04:21:45Z) - Annotators with Attitudes: How Annotator Beliefs And Identities Bias
Toxic Language Detection [75.54119209776894]
We investigate the effect of annotator identities (who) and beliefs (why) on toxic language annotations.
We consider posts with three characteristics: anti-Black language, African American English dialect, and vulgarity.
Our results show strong associations between annotator identity and beliefs and their ratings of toxicity.
arXiv Detail & Related papers (2021-11-15T18:58:20Z) - #ContextMatters: Advantages and Limitations of Using Machine Learning to
Support Women in Politics [0.15749416770494704]
ParityBOT was deployed across elections in Canada, the United States and New Zealand.
It was used to analyse and classify more than 12 million tweets directed at women candidates and counter toxic tweets with supportive ones.
We examine the rate of false negatives, where ParityBOT failed to pick up on insults directed at specific high profile women.
arXiv Detail & Related papers (2021-09-30T22:55:49Z) - Robustness Disparities in Commercial Face Detection [72.25318723264215]
We present the first of its kind detailed benchmark of the robustness of three such systems: Amazon Rekognition, Microsoft Azure, and Google Cloud Platform.
We generally find that photos of individuals who are older, masculine presenting, of darker skin type, or have dim lighting are more susceptible to errors than their counterparts in other identities.
arXiv Detail & Related papers (2021-08-27T21:37:16Z) - SWSR: A Chinese Dataset and Lexicon for Online Sexism Detection [9.443571652110663]
We propose the first Chinese sexism dataset -- Sina Weibo Sexism Review (SWSR) dataset -- and a large Chinese lexicon SexHateLex.
SWSR dataset provides labels at different levels of granularity including (i) sexism or non-sexism, (ii) sexism category and (iii) target type.
We conduct experiments for the three sexism classification tasks making use of state-of-the-art machine learning models.
arXiv Detail & Related papers (2021-08-06T12:06:40Z) - Towards Understanding and Mitigating Social Biases in Language Models [107.82654101403264]
Large-scale pretrained language models (LMs) can be potentially dangerous in manifesting undesirable representational biases.
We propose steps towards mitigating social biases during text generation.
Our empirical results and human evaluation demonstrate effectiveness in mitigating bias while retaining crucial contextual information.
arXiv Detail & Related papers (2021-06-24T17:52:43Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - Automatic Detection of Sexist Statements Commonly Used at the Workplace [0.9790236766474201]
We present a dataset of sexist statements that are more likely to be said in the workplace.
We also present a deep learning model that can achieve state-of-the art results.
arXiv Detail & Related papers (2020-07-08T15:14:29Z) - Echo Chambers on Social Media: A comparative analysis [64.2256216637683]
We introduce an operational definition of echo chambers and perform a massive comparative analysis on 1B pieces of contents produced by 1M users on four social media platforms.
We infer the leaning of users about controversial topics and reconstruct their interaction networks by analyzing different features.
We find support for the hypothesis that platforms implementing news feed algorithms like Facebook may elicit the emergence of echo-chambers.
arXiv Detail & Related papers (2020-04-20T20:00:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.