AI Feedback Enhances Community-Based Content Moderation through Engagement with Counterarguments
- URL: http://arxiv.org/abs/2507.08110v3
- Date: Mon, 06 Oct 2025 11:44:57 GMT
- Title: AI Feedback Enhances Community-Based Content Moderation through Engagement with Counterarguments
- Authors: Saeedeh Mohammadi, Taha Yasseri,
- Abstract summary: This study explores an AI-assisted hybrid moderation framework in which participants receive AI-generated feedback on their notes.<n>The results show that incorporating feedback improves the quality of notes, with the most substantial gains resulting from argumentative feedback.<n>The research contributes to ongoing discussions about AI's role in political content moderation.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Today, social media platforms are significant sources of news and political communication, but their role in spreading misinformation has raised significant concerns. In response, these platforms have implemented various content moderation strategies. One such method, Community Notes (formerly Birdwatch) on X (formerly Twitter), relies on crowdsourced fact-checking and has gained traction. However, it faces challenges such as partisan bias and delays in verification. This study explores an AI-assisted hybrid moderation framework in which participants receive AI-generated feedback, supportive, neutral, or argumentative, on their notes and are asked to revise them accordingly. The results show that incorporating feedback improves the quality of notes, with the most substantial gains resulting from argumentative feedback. This underscores the value of diverse perspectives and direct engagement in human-AI collective intelligence. The research contributes to ongoing discussions about AI's role in political content moderation, highlighting the potential of generative AI and the importance of informed design.
Related papers
- Industrialized Deception: The Collateral Effects of LLM-Generated Misinformation on Digital Ecosystems [47.03825808787752]
This paper transitions from literature review to practical countermeasures.<n>We report on improved AI-generated content through Large Language Models (LLMs) and multimodal systems.<n>We discuss mitigation strategies including LLM-based detection, inoculation approaches, and the dual-use nature of generative AI.
arXiv Detail & Related papers (2026-01-29T16:42:22Z) - The Benefit of Collective Intelligence in Community-Based Content Moderation is Limited by Overt Political Signalling [0.0]
We show that community-based content moderation systems can allow political biases to influence the development of notes and the rating processes.<n>We conduct an online experiment in which participants jointly authored notes on political posts.<n>We find that politically diverse teams perform better when evaluating Republican posts, while group composition does not affect perceived note quality for Democrat posts.
arXiv Detail & Related papers (2026-01-29T16:23:50Z) - From SERPs to Sound: How Search Engine Result Pages and AI-generated Podcasts Interact to Influence User Attitudes on Controversial Topics [18.17104725797712]
We investigate how search engine result pages (SERPs) and AI-generated podcasts interact to shape user opinions.<n>A majority of users in our study corresponded to attitude change outcomes, and we found an effect of sequence on attitude change.<n>Our results further revealed a role of viewpoint bias and the degree of topic controversiality in shaping attitude change, although we found no effect of individual moderators.
arXiv Detail & Related papers (2026-01-16T13:31:11Z) - AI summaries in online search influence users' attitudes [3.459756369056329]
This study examined how AI-generated summaries affect how users think about different issues.<n>Users perceived the AI summaries more useful when it emphasized health harms versus benefits.<n>These findings suggest that AI-generated search summaries can significantly shape public perceptions.
arXiv Detail & Related papers (2025-11-27T23:45:19Z) - AI Credibility Signals Outrank Institutions and Engagement in Shaping News Perception on Social Media [4.197003225775791]
We present a large-scale mixed-design experiment investigating how AI-generated credibility scores affect user perception of political news.<n>Our results reveal that AI feedback significantly moderates partisan bias and institutional distrust, surpassing traditional engagement signals such as likes and shares.
arXiv Detail & Related papers (2025-11-04T08:46:54Z) - CoCoNUTS: Concentrating on Content while Neglecting Uninformative Textual Styles for AI-Generated Peer Review Detection [60.52240468810558]
We introduce CoCoNUTS, a content-oriented benchmark built upon a fine-grained dataset of AI-generated peer reviews.<n>We also develop CoCoDet, an AI review detector via a multi-task learning framework, to achieve more accurate and robust detection of AI involvement in review content.
arXiv Detail & Related papers (2025-08-28T06:03:11Z) - The AI Imperative: Scaling High-Quality Peer Review in Machine Learning [49.87236114682497]
We argue that AI-assisted peer review must become an urgent research and infrastructure priority.<n>We propose specific roles for AI in enhancing factual verification, guiding reviewer performance, assisting authors in quality improvement, and supporting ACs in decision-making.
arXiv Detail & Related papers (2025-06-09T18:37:14Z) - When Models Know More Than They Can Explain: Quantifying Knowledge Transfer in Human-AI Collaboration [79.69935257008467]
We introduce Knowledge Integration and Transfer Evaluation (KITE), a conceptual and experimental framework for Human-AI knowledge transfer capabilities.<n>We conduct the first large-scale human study (N=118) explicitly designed to measure it.<n>In our two-phase setup, humans first ideate with an AI on problem-solving strategies, then independently implement solutions, isolating model explanations' influence on human understanding.
arXiv Detail & Related papers (2025-06-05T20:48:16Z) - Community Moderation and the New Epistemology of Fact Checking on Social Media [124.26693978503339]
Social media platforms have traditionally relied on independent fact-checking organizations to identify and flag misleading content.<n>X (formerly Twitter) and Meta have shifted towards community-driven content moderation by launching their own versions of crowd-sourced fact-checking.<n>We examine the current approaches to misinformation detection across major platforms, explore the emerging role of community-driven moderation, and critically evaluate both the promises and challenges of crowd-checking at scale.
arXiv Detail & Related papers (2025-05-26T14:50:18Z) - The impact of AI and peer feedback on research writing skills: a study using the CGScholar platform among Kazakhstani scholars [0.0]
This research studies the impact of AI and peer feedback on the academic writing development of Kazakhstani scholars using the CGScholar platform.<n>The study aimed to find out how familiarity with AI tools and peer feedback processes impacts participants' openness to incorporating feedback into their academic writing.
arXiv Detail & Related papers (2025-03-05T04:34:25Z) - Users Favor LLM-Generated Content -- Until They Know It's AI [0.0]
We investigate how individuals evaluate human and large langue models generated responses to popular questions when the source of the content is either concealed or disclosed.<n>Our findings indicate that, overall, participants tend to prefer AI-generated responses.<n>When the AI origin is revealed, this preference diminishes significantly, suggesting that evaluative judgments are influenced by the disclosure of the response's provenance.
arXiv Detail & Related papers (2025-02-23T11:14:02Z) - Challenges and Future Directions of Data-Centric AI Alignment [22.165745901158804]
Current alignment methods primarily focus on designing algorithms and loss functions but often underestimate the crucial role of data.<n>This paper advocates for a shift towards data-centric AI alignment, emphasizing the need to enhance the quality and representativeness of data used in aligning AI systems.
arXiv Detail & Related papers (2024-10-02T19:03:42Z) - A Survey of Stance Detection on Social Media: New Directions and Perspectives [50.27382951812502]
stance detection has emerged as a crucial subfield within affective computing.
Recent years have seen a surge of research interest in developing effective stance detection methods.
This paper provides a comprehensive survey of stance detection techniques on social media.
arXiv Detail & Related papers (2024-09-24T03:06:25Z) - HearHere: Mitigating Echo Chambers in News Consumption through an
AI-based Web System [23.289938642423298]
We present HearHere, an AI-based web system designed to help users accommodate information and opinions from diverse perspectives.
Our findings highlight the importance of providing political stance information and quantifying users' political status as a means to mitigate political polarization.
arXiv Detail & Related papers (2024-02-28T10:37:14Z) - The effect of source disclosure on evaluation of AI-generated messages:
A two-part study [0.0]
We examined the influence of source disclosure on people's evaluation of AI-generated health prevention messages.
We found that source disclosure significantly impacted the evaluation of the messages but did not significantly alter message rankings.
For those with moderate levels of negative attitudes towards AI, source disclosure decreased the preference for AI-generated messages.
arXiv Detail & Related papers (2023-11-27T05:20:47Z) - Do You Trust ChatGPT? -- Perceived Credibility of Human and AI-Generated
Content [0.8602553195689513]
This paper examines how individuals perceive the credibility of content originating from human authors versus content generated by large language models.
Surprisingly, our results demonstrate that regardless of the user interface presentation, participants tend to attribute similar levels of credibility.
Participants also do not report any different perceptions of competence and trustworthiness between human and AI-generated content.
arXiv Detail & Related papers (2023-09-05T18:29:29Z) - Fairness And Bias in Artificial Intelligence: A Brief Survey of Sources,
Impacts, And Mitigation Strategies [11.323961700172175]
This survey paper offers a succinct, comprehensive overview of fairness and bias in AI.
We review sources of bias, such as data, algorithm, and human decision biases.
We assess the societal impact of biased AI systems, focusing on the perpetuation of inequalities and the reinforcement of harmful stereotypes.
arXiv Detail & Related papers (2023-04-16T03:23:55Z) - The Role of AI in Drug Discovery: Challenges, Opportunities, and
Strategies [97.5153823429076]
The benefits, challenges and drawbacks of AI in this field are reviewed.
The use of data augmentation, explainable AI, and the integration of AI with traditional experimental methods are also discussed.
arXiv Detail & Related papers (2022-12-08T23:23:39Z) - Aggression and "hate speech" in communication of media users: analysis
of control capabilities [50.591267188664666]
Authors studied the possibilities of mutual influence of users in new media.
They found a high level of aggression and hate speech when discussing an urgent social problem - measures for COVID-19 fighting.
Results can be useful for developing media content in a modern digital environment.
arXiv Detail & Related papers (2022-08-25T15:53:32Z) - Machine Learning Explanations to Prevent Overtrust in Fake News
Detection [64.46876057393703]
This research investigates the effects of an Explainable AI assistant embedded in news review platforms for combating the propagation of fake news.
We design a news reviewing and sharing interface, create a dataset of news stories, and train four interpretable fake news detection algorithms.
For a deeper understanding of Explainable AI systems, we discuss interactions between user engagement, mental model, trust, and performance measures in the process of explaining.
arXiv Detail & Related papers (2020-07-24T05:42:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.