Do You See What I See? Capabilities and Limits of Automated Multimedia
Content Analysis
- URL: http://arxiv.org/abs/2201.11105v1
- Date: Wed, 15 Dec 2021 22:42:00 GMT
- Title: Do You See What I See? Capabilities and Limits of Automated Multimedia
Content Analysis
- Authors: Carey Shenkman, Dhanaraj Thakur, Emma Llans\'o
- Abstract summary: This paper explains the capabilities and limitations of automated content analysis tools.
It focuses on two main categories of tools: matching models and computer prediction models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The ever-increasing amount of user-generated content online has led, in
recent years, to an expansion in research and investment in automated content
analysis tools. Scrutiny of automated content analysis has accelerated during
the COVID-19 pandemic, as social networking services have placed a greater
reliance on these tools due to concerns about health risks to their moderation
staff from in-person work. At the same time, there are important policy debates
around the world about how to improve content moderation while protecting free
expression and privacy. In order to advance these debates, we need to
understand the potential role of automated content analysis tools.
This paper explains the capabilities and limitations of tools for analyzing
online multimedia content and highlights the potential risks of using these
tools at scale without accounting for their limitations. It focuses on two main
categories of tools: matching models and computer prediction models. Matching
models include cryptographic and perceptual hashing, which compare
user-generated content with existing and known content. Predictive models
(including computer vision and computer audition) are machine learning
techniques that aim to identify characteristics of new or previously unknown
content.
Related papers
- A Survey of AI-Generated Video Evaluation [9.100408575312281]
This survey identifies the emerging field of AI-Generated Video Evaluation (AIGVE)
We advocate for the development of more robust and nuanced evaluation frameworks that can handle the complexities of video content.
This survey aims to establish a foundational knowledge base for both researchers from academia and practitioners from the industry.
arXiv Detail & Related papers (2024-10-24T23:08:39Z) - Modeling offensive content detection for TikTok [0.0]
This research undertakes the collection and analysis of TikTok data containing offensive content.
It builds a series of machine learning and deep learning models for offensive content detection.
arXiv Detail & Related papers (2024-08-29T18:47:41Z) - A Survey on Personalized Content Synthesis with Diffusion Models [57.01364199734464]
PCS aims to customize the subject of interest to specific user-defined prompts.
Over the past two years, more than 150 methods have been proposed.
This paper offers a comprehensive survey of PCS, with a particular focus on the diffusion models.
arXiv Detail & Related papers (2024-05-09T04:36:04Z) - Recent Advances in Hate Speech Moderation: Multimodality and the Role of Large Models [52.24001776263608]
This comprehensive survey delves into the recent strides in HS moderation.
We highlight the burgeoning role of large language models (LLMs) and large multimodal models (LMMs)
We identify existing gaps in research, particularly in the context of underrepresented languages and cultures.
arXiv Detail & Related papers (2024-01-30T03:51:44Z) - GPT-4V(ision) as A Social Media Analysis Engine [77.23394183063238]
This paper explores GPT-4V's capabilities for social multimedia analysis.
We select five representative tasks, including sentiment analysis, hate speech detection, fake news identification, demographic inference, and political ideology detection.
GPT-4V demonstrates remarkable efficacy in these tasks, showcasing strengths such as joint understanding of image-text pairs, contextual and cultural awareness, and extensive commonsense knowledge.
arXiv Detail & Related papers (2023-11-13T18:36:50Z) - A Survey on Detection of LLMs-Generated Content [97.87912800179531]
The ability to detect LLMs-generated content has become of paramount importance.
We aim to provide a detailed overview of existing detection strategies and benchmarks.
We also posit the necessity for a multi-faceted approach to defend against various attacks.
arXiv Detail & Related papers (2023-10-24T09:10:26Z) - Countering Malicious Content Moderation Evasion in Online Social
Networks: Simulation and Detection of Word Camouflage [64.78260098263489]
Twisting and camouflaging keywords are among the most used techniques to evade platform content moderation systems.
This article contributes significantly to countering malicious information by developing multilingual tools to simulate and detect new methods of evasion of content.
arXiv Detail & Related papers (2022-12-27T16:08:49Z) - Time-aware topic identification in social media with pre-trained
language models: A case study of electric vehicles [1.2891210250935146]
We propose a time-aware topic identification approach with pre-trained language models.
The proposed approach consists of two stages: the dynamics-focused function for tracking time-varying topics with language models and the emergence-scoring function to examine future promising topics.
arXiv Detail & Related papers (2022-10-11T04:50:10Z) - A Survey of Task-Based Machine Learning Content Extraction Services for
VIDINT [0.0]
Video intelligence (VIDINT) data has become a critical intelligence source in the past decade.
The need for AI-based analytics and automation tools to extract and structure content from video has quickly become a priority for organizations.
This paper reviews and compares products, software resources and video analytics capabilities based on tasks relevant to extracting information from video with machine learning techniques.
arXiv Detail & Related papers (2022-07-09T00:02:08Z) - Automatic Image Content Extraction: Operationalizing Machine Learning in
Humanistic Photographic Studies of Large Visual Archives [81.88384269259706]
We introduce Automatic Image Content Extraction framework for machine learning-based search and analysis of large image archives.
The proposed framework can be applied in several domains in humanities and social sciences.
arXiv Detail & Related papers (2022-04-05T12:19:24Z) - Ownership and Creativity in Generative Models [0.0]
Machine learning generated content such as image artworks, textual poems and music become prominent in recent years.
Because these tools are data-driven, they are inherently different from the traditional creative tools.
Who may own the content that is generated by these tools?
arXiv Detail & Related papers (2021-12-02T18:59:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.