An Order-Complexity Model for Aesthetic Quality Assessment of Homophony
Music Performance
- URL: http://arxiv.org/abs/2304.11521v1
- Date: Sun, 23 Apr 2023 03:02:24 GMT
- Title: An Order-Complexity Model for Aesthetic Quality Assessment of Homophony
Music Performance
- Authors: Xin Jin, Wu Zhou, Jinyu Wang, Duo Xu, Yiqing Rong, Jialin Sun
- Abstract summary: subjective evaluation is still a ultimate method of music aesthetics research.
The music performance generated by AI is still mechanical, monotonous and lacking in beauty.
This paper uses Birkhoff's aesthetic measure to propose a method of objective measurement of beauty.
- Score: 8.751312368054016
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although computational aesthetics evaluation has made certain achievements in
many fields, its research of music performance remains to be explored. At
present, subjective evaluation is still a ultimate method of music aesthetics
research, but it will consume a lot of human and material resources. In
addition, the music performance generated by AI is still mechanical, monotonous
and lacking in beauty. In order to guide the generation task of AI music
performance, and to improve the performance effect of human performers, this
paper uses Birkhoff's aesthetic measure to propose a method of objective
measurement of beauty. The main contributions of this paper are as follows:
Firstly, we put forward an objective aesthetic evaluation method to measure the
music performance aesthetic; Secondly, we propose 10 basic music features and 4
aesthetic music features. Experiments show that our method performs well on
performance assessment.
Related papers
- Towards Explainable and Interpretable Musical Difficulty Estimation: A Parameter-efficient Approach [49.2787113554916]
Estimating music piece difficulty is important for organizing educational music collections.
Our work employs explainable descriptors for difficulty estimation in symbolic music representations.
Our approach, evaluated in piano repertoire categorized in 9 classes, achieved 41.4% accuracy independently, with a mean squared error (MSE) of 1.7.
arXiv Detail & Related papers (2024-08-01T11:23:42Z) - MeLFusion: Synthesizing Music from Image and Language Cues using Diffusion Models [57.47799823804519]
We are inspired by how musicians compose music not just from a movie script, but also through visualizations.
We propose MeLFusion, a model that can effectively use cues from a textual description and the corresponding image to synthesize music.
Our exhaustive experimental evaluation suggests that adding visual information to the music synthesis pipeline significantly improves the quality of generated music.
arXiv Detail & Related papers (2024-06-07T06:38:59Z) - An Order-Complexity Aesthetic Assessment Model for Aesthetic-aware Music
Recommendation [20.164044758068634]
subjective evaluation is still the most effective form of evaluating artistic works.
While compared to music produced by humans, AI generated music still sounds mechanical, monotonous, and lacks aesthetic appeal.
We use Birkhoff's aesthetic measure to design a aesthetic model, objectively measuring the aesthetic beauty of music, and form a recommendation list according to the aesthetic feeling of music.
arXiv Detail & Related papers (2024-02-13T09:03:03Z) - Evaluating Co-Creativity using Total Information Flow [6.3289703660543495]
Co-creativity in music refers to two or more musicians or musical agents interacting with one another by composing or improvising music.
We propose a method to compute the information flow using pre-trained generative models as entropy estimators.
arXiv Detail & Related papers (2024-02-09T22:15:39Z) - MusicRL: Aligning Music Generation to Human Preferences [62.44903326718772]
MusicRL is the first music generation system finetuned from human feedback.
We deploy MusicLM to users and collect a substantial dataset comprising 300,000 pairwise preferences.
We train MusicRL-U, the first text-to-music model that incorporates human feedback at scale.
arXiv Detail & Related papers (2024-02-06T18:36:52Z) - A Comprehensive Survey for Evaluation Methodologies of AI-Generated
Music [14.453416870193072]
This study aims to comprehensively evaluate the subjective, objective, and combined methodologies for assessing AI-generated music.
Ultimately, this study provides a valuable reference for unifying generative AI in the field of music evaluation.
arXiv Detail & Related papers (2023-08-26T02:44:33Z) - An Order-Complexity Model for Aesthetic Quality Assessment of Symbolic
Homophony Music Scores [8.751312368054016]
The quality of music score generated by AI is relatively poor compared with that created by human composers.
This paper proposes an objective quantitative evaluation method for homophony music score aesthetic quality assessment.
arXiv Detail & Related papers (2023-01-14T12:30:16Z) - Musical Prosody-Driven Emotion Classification: Interpreting Vocalists
Portrayal of Emotions Through Machine Learning [0.0]
The role of musical prosody remains under-explored despite several studies demonstrating a strong connection between prosody and emotion.
In this study, we restrict the input of traditional machine learning algorithms to the features of musical prosody.
We utilize a methodology for individual data collection from vocalists, and personal ground truth labeling by the artist themselves.
arXiv Detail & Related papers (2021-06-04T15:40:19Z) - Music Gesture for Visual Sound Separation [121.36275456396075]
"Music Gesture" is a keypoint-based structured representation to explicitly model the body and finger movements of musicians when they perform music.
We first adopt a context-aware graph network to integrate visual semantic context with body dynamics, and then apply an audio-visual fusion model to associate body movements with the corresponding audio signals.
arXiv Detail & Related papers (2020-04-20T17:53:46Z) - Audio Impairment Recognition Using a Correlation-Based Feature
Representation [85.08880949780894]
We propose a new representation of hand-crafted features that is based on the correlation of feature pairs.
We show superior performance in terms of compact feature dimensionality and improved computational speed in the test stage.
arXiv Detail & Related papers (2020-03-22T13:34:37Z) - Multi-Modal Music Information Retrieval: Augmenting Audio-Analysis with
Visual Computing for Improved Music Video Analysis [91.3755431537592]
This thesis combines audio-analysis with computer vision to approach Music Information Retrieval (MIR) tasks from a multi-modal perspective.
The main hypothesis of this work is based on the observation that certain expressive categories such as genre or theme can be recognized on the basis of the visual content alone.
The experiments are conducted for three MIR tasks Artist Identification, Music Genre Classification and Cross-Genre Classification.
arXiv Detail & Related papers (2020-02-01T17:57:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.