Detecting Cultural Differences in News Video Thumbnails via Computational Aesthetics
- URL: http://arxiv.org/abs/2505.21912v1
- Date: Wed, 28 May 2025 02:58:41 GMT
- Title: Detecting Cultural Differences in News Video Thumbnails via Computational Aesthetics
- Authors: Marvin Limpijankit, John Kender,
- Abstract summary: We propose a two-step approach for detecting differences in the style of images across sources of differing cultural affinity.<n>We test this approach on 2,400 YouTube video thumbnails taken equally from two U.S. and two Chinese YouTube channels.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a two-step approach for detecting differences in the style of images across sources of differing cultural affinity, where images are first clustered into finer visual themes based on content before their aesthetic features are compared. We test this approach on 2,400 YouTube video thumbnails taken equally from two U.S. and two Chinese YouTube channels, and relating equally to COVID-19 and the Ukraine conflict. Our results suggest that while Chinese thumbnails are less formal and more candid, U.S. channels tend to use more deliberate, proper photographs as thumbnails. In particular, U.S. thumbnails are less colorful, more saturated, darker, more finely detailed, less symmetric, sparser, less varied, and more up close and personal than Chinese thumbnails. We suggest that most of these differences reflect cultural preferences, and that our methods and observations can serve as a baseline against which suspected visual propaganda can be computed and compared.
Related papers
- Towards Micro-video Thumbnail Selection via a Multi-label
Visual-semantic Embedding Model [0.0]
The thumbnail, as the first sight of a micro-video, plays a pivotal role in attracting users to click and watch.
We present a multi-label visual-semantic embedding model to estimate the similarity between the pair of each frame and the popular topics that users are interested in.
We fuse the visual representation score and the popularity score of each frame to select the attractive thumbnail for the given micro-video.
arXiv Detail & Related papers (2022-02-07T04:15:26Z) - Multi-domain Unsupervised Image-to-Image Translation with Appearance
Adaptive Convolution [62.4972011636884]
We propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework.
We exploit the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance.
We show that the proposed method produces visually diverse and plausible results in multiple domains compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-02-06T14:12:34Z) - Understanding and Evaluating Racial Biases in Image Captioning [18.184279793253634]
We study bias propagation pathways within image captioning, focusing specifically on the COCO dataset.
We demonstrate differences in caption performance, sentiment, and word choice between images of lighter versus darker-skinned people.
arXiv Detail & Related papers (2021-06-16T01:07:24Z) - Broaden Your Views for Self-Supervised Video Learning [97.52216510672251]
We introduce BraVe, a self-supervised learning framework for video.
In BraVe, one of the views has access to a narrow temporal window of the video while the other view has a broad access to the video content.
We demonstrate that BraVe achieves state-of-the-art results in self-supervised representation learning on standard video and audio classification benchmarks.
arXiv Detail & Related papers (2021-03-30T17:58:46Z) - Thumbnail: A Novel Data Augmentation for Convolutional Neural Network [6.066543113636522]
We get a generated image by reducing an image to a certain size, which is called as the thumbnail, and pasting it in the random position of the original image.
The generated image retains most of the original image information but also has the global information in the thumbnail.
We find that the idea of thumbnail can be perfectly integrated with Mixed Sample Data Augmentation.
arXiv Detail & Related papers (2021-03-09T10:45:55Z) - L2C: Describing Visual Differences Needs Semantic Understanding of
Individuals [65.87728481187625]
We introduce a Learning-to-Compare model, which learns to understand the semantic structures of two images and compare them while learning to describe each one.
We demonstrate that L2C benefits from a comparison between explicit semantic representations and single-image captions, and generalizes better on the new testing image pairs.
arXiv Detail & Related papers (2021-02-03T03:44:42Z) - Focus Longer to See Better:Recursively Refined Attention for
Fine-Grained Image Classification [148.4492675737644]
Deep Neural Network has shown great strides in the coarse-grained image classification task.
In this paper, we try to focus on these marginal differences to extract more representative features.
Our network repetitively focuses on parts of images to spot small discriminative parts among the classes.
arXiv Detail & Related papers (2020-05-22T03:14:18Z) - Watching the World Go By: Representation Learning from Unlabeled Videos [78.22211989028585]
Recent single image unsupervised representation learning techniques show remarkable success on a variety of tasks.
In this paper, we argue that videos offer this natural augmentation for free.
We propose Video Noise Contrastive Estimation, a method for using unlabeled video to learn strong, transferable single image representations.
arXiv Detail & Related papers (2020-03-18T00:07:21Z) - Mi YouTube es Su YouTube? Analyzing the Cultures using YouTube
Thumbnails of Popular Videos [98.87558262467257]
This study explores culture preferences among countries using the thumbnails of YouTube trending videos.
Experimental results indicate that the users from similar cultures shares interests in watching similar videos on YouTube.
arXiv Detail & Related papers (2020-01-27T20:15:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.