Analysing the Public Discourse around OpenAI's Text-To-Video Model 'Sora' using Topic Modeling
- URL: http://arxiv.org/abs/2407.13071v1
- Date: Thu, 30 May 2024 01:55:30 GMT
- Title: Analysing the Public Discourse around OpenAI's Text-To-Video Model 'Sora' using Topic Modeling
- Authors: Vatsal Vinay Parikh,
- Abstract summary: This study aims to uncover the dominant themes and narratives surrounding Sora by conducting topic modeling analysis on a corpus of 1,827 Reddit comments.
The comments were collected over a two-month period following Sora's announcement in February 2024.
The results highlight prominent narratives around Sora's potential impact on industries and employment, public sentiment and ethical concerns, creative applications, and use cases in the media and entertainment sectors.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: The recent introduction of OpenAI's text-to-video model Sora has sparked widespread public discourse across online communities. This study aims to uncover the dominant themes and narratives surrounding Sora by conducting topic modeling analysis on a corpus of 1,827 Reddit comments from five relevant subreddits (r/OpenAI, r/technology, r/singularity, r/vfx, and r/ChatGPT). The comments were collected over a two-month period following Sora's announcement in February 2024. After preprocessing the data, Latent Dirichlet Allocation (LDA) was employed to extract four key topics: 1) AI Impact and Trends in Sora Discussions, 2) Public Opinion and Concerns about Sora, 3) Artistic Expression and Video Creation with Sora, and 4) Sora's Applications in Media and Entertainment. Visualizations including word clouds, bar charts, and t-SNE clustering provided insights into the importance of topic keywords and the distribution of comments across topics. The results highlight prominent narratives around Sora's potential impact on industries and employment, public sentiment and ethical concerns, creative applications, and use cases in the media and entertainment sectors. While limited to Reddit data within a specific timeframe, this study offers a framework for understanding public perceptions of emerging generative AI technologies through online discourse analysis.
Related papers
- HOTVCOM: Generating Buzzworthy Comments for Videos [49.39846630199698]
This study introduces textscHotVCom, the largest Chinese video hot-comment dataset, comprising 94k diverse videos and 137 million comments.
We also present the textttComHeat framework, which synergistically integrates visual, auditory, and textual data to generate influential hot-comments on the Chinese video dataset.
arXiv Detail & Related papers (2024-09-23T16:45:13Z) - Mapping AI Ethics Narratives: Evidence from Twitter Discourse Between 2015 and 2022 [6.518657832967228]
Twitter is selected in this paper to serve as an online public sphere for exploring discourse on AI ethics.
A research framework is proposed to demonstrate how to transform AI ethics-related discourse on Twitter into coherent and readable narratives.
arXiv Detail & Related papers (2024-06-20T09:08:44Z) - From Sora What We Can See: A Survey of Text-to-Video Generation [10.204414498390575]
Sora, developed by OpenAI, is capable of minute-level world-simulative abilities.
Despite its notable successes, Sora still encounters various obstacles that need to be resolved.
arXiv Detail & Related papers (2024-05-17T10:09:09Z) - "Sora is Incredible and Scary": Emerging Governance Challenges of Text-to-Video Generative AI Models [1.4999444543328293]
We report a qualitative social media analysis aiming to uncover people's perceived impact of and concerns about Sora's integration.
We found that people were most concerned about Sora's impact on content creation-related industries.
Potential regulatory solutions included law-enforced labeling of AI content and AI literacy education for the public.
arXiv Detail & Related papers (2024-04-10T02:03:59Z) - Sora OpenAI's Prelude: Social Media Perspectives on Sora OpenAI and the Future of AI Video Generation [30.556463355261695]
This study investigates the public's perception of Sora OpenAI, a pioneering Gen-AI video generation tool, via social media discussions on Reddit.
The analysis forecasts positive shifts in content creation, predicting that Sora will democratize video marketing and innovate game development.
There are concerns about deepfakes and the potential for disinformation, underscoring the need for strategies to address disinformation and bias.
arXiv Detail & Related papers (2024-03-02T00:16:22Z) - Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models [59.54172719450617]
Sora is a text-to-video generative AI model, released by OpenAI in February 2024.
This paper presents a review of the model's background, related technologies, applications, remaining challenges, and future directions.
arXiv Detail & Related papers (2024-02-27T03:30:58Z) - Sentiment-enhanced Graph-based Sarcasm Explanation in Dialogue [67.09698638709065]
We propose a novel sEntiment-enhanceD Graph-based multimodal sarcasm Explanation framework, named EDGE.
In particular, we first propose a lexicon-guided utterance sentiment inference module, where a utterance sentiment refinement strategy is devised.
We then develop a module named Joint Cross Attention-based Sentiment Inference (JCA-SI) by extending the multimodal sentiment analysis model JCA to derive the joint sentiment label for each video-audio clip.
arXiv Detail & Related papers (2024-02-06T03:14:46Z) - InfoPattern: Unveiling Information Propagation Patterns in Social Media [59.67008841974645]
InfoPattern centers on the interplay between language and human ideology.
The demo is capable of: (1) red teaming to simulate adversary responses from opposite ideology communities; (2) stance detection to identify the underlying political sentiments in each message; (3) information propagation graph discovery to reveal the evolution of claims across various communities over time.
arXiv Detail & Related papers (2023-11-27T09:12:35Z) - Affective Image Content Analysis: Two Decades Review and New
Perspectives [132.889649256384]
We will comprehensively review the development of affective image content analysis (AICA) in the recent two decades.
We will focus on the state-of-the-art methods with respect to three main challenges -- the affective gap, perception subjectivity, and label noise and absence.
We discuss some challenges and promising research directions in the future, such as image content and context understanding, group emotion clustering, and viewer-image interaction.
arXiv Detail & Related papers (2021-06-30T15:20:56Z) - "Notic My Speech" -- Blending Speech Patterns With Multimedia [65.91370924641862]
We propose a view-temporal attention mechanism to model both the view dependence and the visemic importance in speech recognition and understanding.
Our proposed method outperformed the existing work by 4.99% in terms of the viseme error rate.
We show that there is a strong correlation between our model's understanding of multi-view speech and the human perception.
arXiv Detail & Related papers (2020-06-12T06:51:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.