Multi-Modal Aesthetic Assessment for MObile Gaming Image
- URL: http://arxiv.org/abs/2101.11700v1
- Date: Wed, 27 Jan 2021 21:48:31 GMT
- Title: Multi-Modal Aesthetic Assessment for MObile Gaming Image
- Authors: Zhenyu Lei, Yejing Xie, Suiyi Ling, Andreas Pastor, Junle Wang,
Patrick Le Callet
- Abstract summary: The proposed model outperforms state-of-the-art aesthetic metrics significantly in predicting four gaming aesthetic dimensions.
Our inclination is to seek and learn the correlations between different aesthetic relevant dimensions to further boost the generalization performance.
- Score: 30.962059154484912
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the proliferation of various gaming technology, services, game styles,
and platforms, multi-dimensional aesthetic assessment of the gaming contents is
becoming more and more important for the gaming industry. Depending on the
diverse needs of diversified game players, game designers, graphical
developers, etc. in particular conditions, multi-modal aesthetic assessment is
required to consider different aesthetic dimensions/perspectives. Since there
are different underlying relationships between different aesthetic dimensions,
e.g., between the `Colorfulness' and `Color Harmony', it could be advantageous
to leverage effective information attached in multiple relevant dimensions. To
this end, we solve this problem via multi-task learning. Our inclination is to
seek and learn the correlations between different aesthetic relevant dimensions
to further boost the generalization performance in predicting all the aesthetic
dimensions. Therefore, the `bottleneck' of obtaining good predictions with
limited labeled data for one individual dimension could be unplugged by
harnessing complementary sources of other dimensions, i.e., augment the
training data indirectly by sharing training information across dimensions.
According to experimental results, the proposed model outperforms
state-of-the-art aesthetic metrics significantly in predicting four gaming
aesthetic dimensions.
Related papers
- When Does Perceptual Alignment Benefit Vision Representations? [76.32336818860965]
We investigate how aligning vision model representations to human perceptual judgments impacts their usability.
We find that aligning models to perceptual judgments yields representations that improve upon the original backbones across many downstream tasks.
Our results suggest that injecting an inductive bias about human perceptual knowledge into vision models can contribute to better representations.
arXiv Detail & Related papers (2024-10-14T17:59:58Z) - Perceptual Similarity for Measuring Decision-Making Style and Policy Diversity in Games [28.289135305943056]
Defining and measuring decision-making styles, also known as playstyles, is crucial in gaming.
We introduce three enhancements to increase accuracy: multiscale analysis with varied state psychology, a perceptual kernel rooted in granularity, and the utilization of the intersection-over-union method for efficient evaluation.
Our findings improve the measurement of end-to-end game analysis and the evolution of artificial intelligence for diverse playstyles.
arXiv Detail & Related papers (2024-08-12T10:55:42Z) - AesExpert: Towards Multi-modality Foundation Model for Image Aesthetics Perception [74.11069437400398]
We develop a corpus-rich aesthetic critique database with 21,904 diverse-sourced images and 88K human natural language feedbacks.
We fine-tune the open-sourced general foundation models, achieving multi-modality Aesthetic Expert models, dubbed AesExpert.
Experiments demonstrate that the proposed AesExpert models deliver significantly better aesthetic perception performances than the state-of-the-art MLLMs.
arXiv Detail & Related papers (2024-04-15T09:56:20Z) - Measuring Diversity of Game Scenarios [15.100151112002235]
We aim to bridge the current gaps in literature and practice, offering insights into effective strategies for measuring and integrating diversity in game scenarios.
This survey not only charts a path for future research in diverse game scenarios but also serves as a handbook for industry practitioners seeking to leverage diversity as a key component of game design and development.
arXiv Detail & Related papers (2024-04-15T07:59:52Z) - UniAR: A Unified model for predicting human Attention and Responses on visual content [12.281060227170792]
We propose UniAR -- a unified model of human attention and preference behavior across diverse visual content.
We train UniAR on diverse public datasets spanning natural images, webpages, and graphic designs, and achieve SOTA performance on multiple benchmarks.
Potential applications include providing instant feedback on the effectiveness of UIs/visual content, and enabling designers and content-creation models to optimize their creation for human-centric improvements.
arXiv Detail & Related papers (2023-12-15T19:57:07Z) - Impressions: Understanding Visual Semiotics and Aesthetic Impact [66.40617566253404]
We present Impressions, a novel dataset through which to investigate the semiotics of images.
We show that existing multimodal image captioning and conditional generation models struggle to simulate plausible human responses to images.
This dataset significantly improves their ability to model impressions and aesthetic evaluations of images through fine-tuning and few-shot adaptation.
arXiv Detail & Related papers (2023-10-27T04:30:18Z) - Towards General Game Representations: Decomposing Games Pixels into
Content and Style [2.570570340104555]
Learning pixel representations of games can benefit artificial intelligence across several downstream tasks.
This paper explores how generalizable pre-trained computer vision encoders can be for such tasks.
We employ a pre-trained Vision Transformer encoder and a decomposition technique based on game genres to obtain separate content and style embeddings.
arXiv Detail & Related papers (2023-07-20T17:53:04Z) - Perceiver: General Perception with Iterative Attention [85.65927856589613]
We introduce the Perceiver - a model that builds upon Transformers.
We show that this architecture performs competitively or beyond strong, specialized models on classification tasks.
It also surpasses state-of-the-art results for all modalities in AudioSet.
arXiv Detail & Related papers (2021-03-04T18:20:50Z) - Individualized Context-Aware Tensor Factorization for Online Games
Predictions [6.602875221541352]
We present the Neural Individualized Context-aware Embeddings (NICE) model for predicting user performance and game outcomes.
Our proposed method identifies individual behavioral differences in different contexts by learning latent representations of users and contexts.
Using a dataset from the MOBA game League of Legends, we demonstrate that our model substantially improves the prediction of winning outcome, individual user performance, and user engagement.
arXiv Detail & Related papers (2021-02-22T20:46:02Z) - Embedded Deep Bilinear Interactive Information and Selective Fusion for
Multi-view Learning [70.67092105994598]
We propose a novel multi-view learning framework to make the multi-view classification better aimed at the above-mentioned two aspects.
In particular, we train different deep neural networks to learn various intra-view representations.
Experiments on six publicly available datasets demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2020-07-13T01:13:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.