Enhancing Visual Sentiment Analysis via Semiotic Isotopy-Guided Dataset Construction
- URL: http://arxiv.org/abs/2512.14665v1
- Date: Tue, 16 Dec 2025 18:26:22 GMT
- Title: Enhancing Visual Sentiment Analysis via Semiotic Isotopy-Guided Dataset Construction
- Authors: Marco Blanchini, Giovanna Maria Dimitri, Benedetta Tondi, Tarcisio Lancioni, Mauro Barni,
- Abstract summary: We show how to create a new larger dataset for Visual Sentiment Analysis.<n>This allows training new models with improved capability to focus on emotionally relevant combinations of image elements.<n> Empirical evaluations show that models trained on a dataset generated with our method consistently outperform those trained on the original data collections.
- Score: 15.18770071829755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual Sentiment Analysis (VSA) is a challenging task due to the vast diversity of emotionally salient images and the inherent difficulty of acquiring sufficient data to capture this variability comprehensively. Key obstacles include building large-scale VSA datasets and developing effective methodologies that enable algorithms to identify emotionally significant elements within an image. These challenges are reflected in the limited generalization performance of VSA algorithms and models when trained and tested across different datasets. Starting from a pool of existing data collections, our approach enables the creation of a new larger dataset that not only contains a wider variety of images than the original ones, but also permits training new models with improved capability to focus on emotionally relevant combinations of image elements. This is achieved through the integration of the semiotic isotopy concept within the dataset creation process, providing deeper insights into the emotional content of images. Empirical evaluations show that models trained on a dataset generated with our method consistently outperform those trained on the original data collections, achieving superior generalization across major VSA benchmarks
Related papers
- Navigating the Challenges of AI-Generated Image Detection in the Wild: What Truly Matters? [9.916527862912941]
We introduce ITW-SM, a new dataset of real and AI-generated images collected from major social media platforms.<n>We identify four key factors that influence AID performance in real-world scenarios.<n>Our modifications result in an average AUC improvement of 26.87% across various AID models under real-world conditions.
arXiv Detail & Related papers (2025-07-14T12:56:55Z) - Cross-Spectral Body Recognition with Side Information Embedding: Benchmarks on LLCM and Analyzing Range-Induced Occlusions on IJB-MDF [51.36007967653781]
Vision Transformers (ViTs) have demonstrated impressive performance across a wide range of biometric tasks, including face and body recognition.<n>In this work, we adapt a ViT model pretrained on visible (VIS) imagery to the challenging problem of cross-spectral body recognition.<n>Building on this idea, we integrate Side Information Embedding (SIE) and examine the impact of encoding domain and camera information to enhance cross-spectral matching.<n>Surprisingly, our results show that encoding only camera information - without explicitly incorporating domain information - achieves state-of-the-art performance on the LLCM dataset.
arXiv Detail & Related papers (2025-06-10T16:20:52Z) - Picking the Cream of the Crop: Visual-Centric Data Selection with Collaborative Agents [62.616106562146776]
We propose a textbfVisual-Centric textbfSelection approach via textbfAgents Collaboration (ViSA)<n>Our approach consists of 1) an image information quantification method via visual agents collaboration to select images with rich visual information, and 2) a visual-centric instruction quality assessment method to select high-quality instruction data related to high-quality images.
arXiv Detail & Related papers (2025-02-27T09:37:30Z) - A Simple Background Augmentation Method for Object Detection with Diffusion Model [53.32935683257045]
In computer vision, it is well-known that a lack of data diversity will impair model performance.
We propose a simple yet effective data augmentation approach by leveraging advancements in generative models.
Background augmentation, in particular, significantly improves the models' robustness and generalization capabilities.
arXiv Detail & Related papers (2024-08-01T07:40:00Z) - SDFD: Building a Versatile Synthetic Face Image Dataset with Diverse Attributes [14.966767182001755]
We propose a methodology for generating synthetic face image datasets that capture a broader spectrum of facial diversity.
Specifically, our approach integrates demographics and biometrics but also non-permanent traits like make-up, hairstyle, and accessories.
These prompts guide a state-of-the-art text-to-image model in generating a comprehensive dataset of high-quality realistic images.
arXiv Detail & Related papers (2024-04-26T08:51:31Z) - Rotated Multi-Scale Interaction Network for Referring Remote Sensing Image Segmentation [63.15257949821558]
Referring Remote Sensing Image (RRSIS) is a new challenge that combines computer vision and natural language processing.
Traditional Referring Image (RIS) approaches have been impeded by the complex spatial scales and orientations found in aerial imagery.
We introduce the Rotated Multi-Scale Interaction Network (RMSIN), an innovative approach designed for the unique demands of RRSIS.
arXiv Detail & Related papers (2023-12-19T08:14:14Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Gradient-Induced Co-Saliency Detection [81.54194063218216]
Co-saliency detection (Co-SOD) aims to segment the common salient foreground in a group of relevant images.
In this paper, inspired by human behavior, we propose a gradient-induced co-saliency detection method.
arXiv Detail & Related papers (2020-04-28T08:40:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.