Listener Modeling and Context-aware Music Recommendation Based on
Country Archetypes
- URL: http://arxiv.org/abs/2009.09935v1
- Date: Fri, 11 Sep 2020 17:59:04 GMT
- Title: Listener Modeling and Context-aware Music Recommendation Based on
Country Archetypes
- Authors: Markus Schedl, Christine Bauer, Wolfgang Reisinger, Dominik Kowald,
Elisabeth Lex
- Abstract summary: Music preferences are strongly shaped by the cultural and socio-economic background of the listener.
We use state-of-the-art unsupervised learning techniques to investigate country profiles of music preferences on the fine-grained level of music tracks.
We propose a context-aware music recommendation system that leverages implicit user feedback.
- Score: 10.19712238203935
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Music preferences are strongly shaped by the cultural and socio-economic
background of the listener, which is reflected, to a considerable extent, in
country-specific music listening profiles. Previous work has already identified
several country-specific differences in the popularity distribution of music
artists listened to. In particular, what constitutes the "music mainstream"
strongly varies between countries. To complement and extend these results, the
article at hand delivers the following major contributions: First, using
state-of-the-art unsupervised learning techniques, we identify and thoroughly
investigate (1) country profiles of music preferences on the fine-grained level
of music tracks (in contrast to earlier work that relied on music preferences
on the artist level) and (2) country archetypes that subsume countries sharing
similar patterns of listening preferences. Second, we formulate four user
models that leverage the user's country information on music preferences. Among
others, we propose a user modeling approach to describe a music listener as a
vector of similarities over the identified country clusters or archetypes.
Third, we propose a context-aware music recommendation system that leverages
implicit user feedback, where context is defined via the four user models. More
precisely, it is a multi-layer generative model based on a variational
autoencoder, in which contextual features can influence recommendations through
a gating mechanism. Fourth, we thoroughly evaluate the proposed recommendation
system and user models on a real-world corpus of more than one billion
listening records of users around the world (out of which we use 369 million in
our experiments) and show its merits vis-a-vis state-of-the-art algorithms that
do not exploit this type of context information.
Related papers
- MuChoMusic: Evaluating Music Understanding in Multimodal Audio-Language Models [11.834712543531756]
MuChoMusic is a benchmark for evaluating music understanding in multimodal language models focused on audio.
It comprises 1,187 multiple-choice questions, all validated by human annotators, on 644 music tracks sourced from two publicly available music datasets.
We evaluate five open-source models and identify several pitfalls, including an over-reliance on the language modality.
arXiv Detail & Related papers (2024-08-02T15:34:05Z) - MeLFusion: Synthesizing Music from Image and Language Cues using Diffusion Models [57.47799823804519]
We are inspired by how musicians compose music not just from a movie script, but also through visualizations.
We propose MeLFusion, a model that can effectively use cues from a textual description and the corresponding image to synthesize music.
Our exhaustive experimental evaluation suggests that adding visual information to the music synthesis pipeline significantly improves the quality of generated music.
arXiv Detail & Related papers (2024-06-07T06:38:59Z) - MusicRL: Aligning Music Generation to Human Preferences [62.44903326718772]
MusicRL is the first music generation system finetuned from human feedback.
We deploy MusicLM to users and collect a substantial dataset comprising 300,000 pairwise preferences.
We train MusicRL-U, the first text-to-music model that incorporates human feedback at scale.
arXiv Detail & Related papers (2024-02-06T18:36:52Z) - Fairness Through Domain Awareness: Mitigating Popularity Bias For Music
Discovery [56.77435520571752]
We explore the intrinsic relationship between music discovery and popularity bias.
We propose a domain-aware, individual fairness-based approach which addresses popularity bias in graph neural network (GNNs) based recommender systems.
Our approach uses individual fairness to reflect a ground truth listening experience, i.e., if two songs sound similar, this similarity should be reflected in their representations.
arXiv Detail & Related papers (2023-08-28T14:12:25Z) - From West to East: Who can understand the music of the others better? [91.78564268397139]
We leverage transfer learning methods to derive insights about similarities between different music cultures.
We use two Western music datasets, two traditional/folk datasets coming from eastern Mediterranean cultures, and two datasets belonging to Indian art music.
Three deep audio embedding models are trained and transferred across domains, including two CNN-based and a Transformer-based architecture, to perform auto-tagging for each target domain dataset.
arXiv Detail & Related papers (2023-07-19T07:29:14Z) - MARBLE: Music Audio Representation Benchmark for Universal Evaluation [79.25065218663458]
We introduce the Music Audio Representation Benchmark for universaL Evaluation, termed MARBLE.
It aims to provide a benchmark for various Music Information Retrieval (MIR) tasks by defining a comprehensive taxonomy with four hierarchy levels, including acoustic, performance, score, and high-level description.
We then establish a unified protocol based on 14 tasks on 8 public-available datasets, providing a fair and standard assessment of representations of all open-sourced pre-trained models developed on music recordings as baselines.
arXiv Detail & Related papers (2023-06-18T12:56:46Z) - A Dataset for Greek Traditional and Folk Music: Lyra [69.07390994897443]
This paper presents a dataset for Greek Traditional and Folk music that includes 1570 pieces, summing in around 80 hours of data.
The dataset incorporates YouTube timestamped links for retrieving audio and video, along with rich metadata information with regards to instrumentation, geography and genre.
arXiv Detail & Related papers (2022-11-21T14:15:43Z) - Follow the guides: disentangling human and algorithmic curation in
online music consumption [1.4506962780822348]
We analyze the complete listening history of about 9k users over one year.
We show that the two types of recommendation offered by music platforms -- algorithmic and editorial -- may drive the consumption of more or less diverse content in opposite directions.
arXiv Detail & Related papers (2021-09-08T20:14:48Z) - Modeling the Music Genre Perception across Language-Bound Cultures [10.223656553455003]
We study the feasibility of obtaining relevant cross-lingual, culture-specific music genre annotations.
We show that unsupervised cross-lingual music genre annotation is feasible with high accuracy.
We introduce a new, domain-dependent cross-lingual corpus to benchmark state of the art multilingual pre-trained embedding models.
arXiv Detail & Related papers (2020-10-13T12:20:32Z) - Time-Aware Music Recommender Systems: Modeling the Evolution of Implicit
User Preferences and User Listening Habits in A Collaborative Filtering
Approach [4.576379639081977]
This paper studies the temporal information regarding when songs are played.
The purpose is to model both the evolution of user preferences in the form of evolving implicit ratings and user listening behavior.
In the collaborative filtering method proposed in this work, daily listening habits are captured in order to characterize users and provide them with more reliable recommendations.
arXiv Detail & Related papers (2020-08-26T08:00:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.