Artist Similarity with Graph Neural Networks
- URL: http://arxiv.org/abs/2107.14541v1
- Date: Fri, 30 Jul 2021 10:44:31 GMT
- Title: Artist Similarity with Graph Neural Networks
- Authors: Filip Korzeniowski, Sergio Oramas, Fabien Gouyon
- Abstract summary: We present a hybrid approach to computing similarity between artists using graph neural networks trained with triplet loss.
The novelty of using a graph neural network architecture is to combine the topology of a graph of artist connections with content features to embed artists into a vector space that encodes similarity.
With 17,673 artists, this is the largest academic artist similarity dataset that includes content-based features to date.
- Score: 1.160208922584163
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artist similarity plays an important role in organizing, understanding, and
subsequently, facilitating discovery in large collections of music. In this
paper, we present a hybrid approach to computing similarity between artists
using graph neural networks trained with triplet loss. The novelty of using a
graph neural network architecture is to combine the topology of a graph of
artist connections with content features to embed artists into a vector space
that encodes similarity. To evaluate the proposed method, we compile the new
OLGA dataset, which contains artist similarities from AllMusic, together with
content features from AcousticBrainz. With 17,673 artists, this is the largest
academic artist similarity dataset that includes content-based features to
date. Moreover, we also showcase the scalability of our approach by
experimenting with a much larger proprietary dataset. Results show the
superiority of the proposed approach over current state-of-the-art methods for
music similarity. Finally, we hope that the OLGA dataset will facilitate
research on data-driven models for artist similarity.
Related papers
- MeLFusion: Synthesizing Music from Image and Language Cues using Diffusion Models [57.47799823804519]
We are inspired by how musicians compose music not just from a movie script, but also through visualizations.
We propose MeLFusion, a model that can effectively use cues from a textual description and the corresponding image to synthesize music.
Our exhaustive experimental evaluation suggests that adding visual information to the music synthesis pipeline significantly improves the quality of generated music.
arXiv Detail & Related papers (2024-06-07T06:38:59Z) - Learning Multimodal Volumetric Features for Large-Scale Neuron Tracing [72.45257414889478]
We aim to reduce human workload by predicting connectivity between over-segmented neuron pieces.
We first construct a dataset, named FlyTracing, that contains millions of pairwise connections of segments expanding the whole fly brain.
We propose a novel connectivity-aware contrastive learning method to generate dense volumetric EM image embedding.
arXiv Detail & Related papers (2024-01-05T19:45:12Z) - Combinatorial music generation model with song structure graph analysis [18.71152526968065]
We construct a graph that uses information such as note sequence and instrument as node features, while the correlation between note sequences acts as the edge feature.
We trained a Graph Neural Network to obtain node representation in the graph, then we use node representation as input of Unet to generate CONLON pianoroll image latent.
arXiv Detail & Related papers (2023-12-24T04:09:30Z) - Representing Videos as Discriminative Sub-graphs for Action Recognition [165.54738402505194]
We introduce a new design of sub-graphs to represent and encode theriminative patterns of each action in the videos.
We present MUlti-scale Sub-Earn Ling (MUSLE) framework that novelly builds space-time graphs and clusters into compact sub-graphs on each scale.
arXiv Detail & Related papers (2022-01-11T16:15:25Z) - Learning Co-segmentation by Segment Swapping for Retrieval and Discovery [67.6609943904996]
The goal of this work is to efficiently identify visually similar patterns from a pair of images.
We generate synthetic training pairs by selecting object segments in an image and copy-pasting them into another image.
We show our approach provides clear improvements for artwork details retrieval on the Brueghel dataset.
arXiv Detail & Related papers (2021-10-29T16:51:16Z) - Cold Start Similar Artists Ranking with Gravity-Inspired Graph
Autoencoders [18.395568778680207]
We model a cold start similar artists ranking problem as a link prediction task in a directed and attributed graph.
Then, we leverage a graph autoencoder architecture to learn node embedding representations from this graph, and to automatically rank the top-k most similar neighbors of new artists.
We empirically show the flexibility and the effectiveness of our framework, by addressing a real-world cold start similar artists ranking problem on a global music streaming service.
arXiv Detail & Related papers (2021-08-02T17:19:47Z) - Graph Neural Networks for Knowledge Enhanced Visual Representation of
Paintings [14.89186519385364]
ArtSAGENet is a novel architecture that integrates Graph Neural Networks (GNNs) and Convolutional Neural Networks (CNNs)
We show that our proposed ArtSAGENet captures and encodes valuable dependencies between the artists and the artworks.
Our findings underline a great potential of integrating visual content and semantics for fine art analysis and curation.
arXiv Detail & Related papers (2021-05-17T23:05:36Z) - Exploiting Heterogeneous Graph Neural Networks with Latent Worker/Task
Correlation Information for Label Aggregation in Crowdsourcing [72.34616482076572]
Crowdsourcing has attracted much attention for its convenience to collect labels from non-expert workers instead of experts.
We propose a novel framework based on graph neural networks for aggregating crowd labels.
arXiv Detail & Related papers (2020-10-25T10:12:37Z) - Detecting Generic Music Features with Single Layer Feedforward Network
using Unsupervised Hebbian Computation [3.8707695363745223]
The authors extract information on such features from a popular open-source music corpus.
They apply unsupervised Hebbian learning techniques on their single-layer neural network using the same dataset.
The unsupervised training algorithm enhances their proposed neural network to achieve an accuracy of 90.36% for successful music feature detection.
arXiv Detail & Related papers (2020-08-31T13:57:31Z) - Content-based Music Similarity with Triplet Networks [21.220806977978853]
We explore the feasibility of using triplet neural networks to embed songs based on content-based music similarity.
Our network is trained using triplets of songs such that two songs by the same artist are embedded closer to one another than to a third song by a different artist.
arXiv Detail & Related papers (2020-08-11T18:10:02Z) - dMelodies: A Music Dataset for Disentanglement Learning [70.90415511736089]
We present a new symbolic music dataset that will help researchers demonstrate the efficacy of their algorithms on diverse domains.
This will also provide a means for evaluating algorithms specifically designed for music.
The dataset is large enough (approx. 1.3 million data points) to train and test deep networks for disentanglement learning.
arXiv Detail & Related papers (2020-07-29T19:20:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.