Attentional networks for music generation
- URL: http://arxiv.org/abs/2002.03854v1
- Date: Thu, 6 Feb 2020 13:26:17 GMT
- Title: Attentional networks for music generation
- Authors: Gullapalli Keerti, A N Vaishnavi, Prerana Mukherjee, A Sree Vidya,
Gattineni Sai Sreenithya, Deeksha Nayab
- Abstract summary: We propose a deep learning based music generation method in order to produce old style music particularly JAZZ with rehashed melodic structures.
Owing to the success in modelling long-term temporal dependencies in sequential data and its success in case of videos, Bi-LSTMs with attention serve as the natural choice and early utilization in music generation.
We validate in our experiments that Bi-LSTMs with attention are able to preserve the richness and technical nuances of the music performed.
- Score: 5.012960295592238
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Realistic music generation has always remained as a challenging problem as it
may lack structure or rationality. In this work, we propose a deep learning
based music generation method in order to produce old style music particularly
JAZZ with rehashed melodic structures utilizing a Bi-directional Long Short
Term Memory (Bi-LSTM) Neural Network with Attention. Owing to the success in
modelling long-term temporal dependencies in sequential data and its success in
case of videos, Bi-LSTMs with attention serve as the natural choice and early
utilization in music generation. We validate in our experiments that Bi-LSTMs
with attention are able to preserve the richness and technical nuances of the
music performed.
Related papers
- MuVi: Video-to-Music Generation with Semantic Alignment and Rhythmic Synchronization [52.498942604622165]
This paper presents MuVi, a framework to generate music that aligns with video content.
MuVi analyzes video content through a specially designed visual adaptor to extract contextually and temporally relevant features.
We show that MuVi demonstrates superior performance in both audio quality and temporal synchronization.
arXiv Detail & Related papers (2024-10-16T18:44:56Z) - MMT-BERT: Chord-aware Symbolic Music Generation Based on Multitrack Music Transformer and MusicBERT [44.204383306879095]
We propose a novel symbolic music representation and Generative Adversarial Network (GAN) framework specially designed for symbolic multitrack music generation.
To build a robust multitrack music generator, we fine-tune a pre-trained MusicBERT model to serve as the discriminator, and incorporate relativistic standard loss.
arXiv Detail & Related papers (2024-09-02T03:18:56Z) - Generating Music with Structure Using Self-Similarity as Attention [0.0]
We propose an attention layer that uses a novel approach applying user-supplied self-similarity to previous time steps.
We demonstrate it in our Similarity Incentivized Generator (SING) system, a deep learning autonomous music generation system.
We train SING on the MAESTRO dataset using a novel variable method, and compare its performance to the same model without the attention mechanism.
arXiv Detail & Related papers (2024-06-21T20:56:12Z) - Music Emotion Prediction Using Recurrent Neural Networks [8.867897390286815]
This study aims to enhance music recommendation systems and support therapeutic interventions by tailoring music to fit listeners' emotional states.
We utilize Russell's Emotion Quadrant to categorize music into four distinct emotional regions and develop models capable of accurately predicting these categories.
Our approach involves extracting a comprehensive set of audio features using Librosa and applying various recurrent neural network architectures, including standard RNNs, Bidirectional RNNs, and Long Short-Term Memory (LSTM) networks.
arXiv Detail & Related papers (2024-05-10T18:03:20Z) - MuPT: A Generative Symbolic Music Pretrained Transformer [56.09299510129221]
We explore the application of Large Language Models (LLMs) to the pre-training of music.
To address the challenges associated with misaligned measures from different tracks during generation, we propose a Synchronized Multi-Track ABC Notation (SMT-ABC Notation)
Our contributions include a series of models capable of handling up to 8192 tokens, covering 90% of the symbolic music data in our training set.
arXiv Detail & Related papers (2024-04-09T15:35:52Z) - GETMusic: Generating Any Music Tracks with a Unified Representation and
Diffusion Framework [58.64512825534638]
Symbolic music generation aims to create musical notes, which can help users compose music.
We introduce a framework known as GETMusic, with GET'' standing for GEnerate music Tracks''
GETScore represents musical notes as tokens and organizes tokens in a 2D structure, with tracks stacked vertically and progressing horizontally over time.
Our proposed representation, coupled with the non-autoregressive generative model, empowers GETMusic to generate music with any arbitrary source-target track combinations.
arXiv Detail & Related papers (2023-05-18T09:53:23Z) - Museformer: Transformer with Fine- and Coarse-Grained Attention for
Music Generation [138.74751744348274]
We propose Museformer, a Transformer with a novel fine- and coarse-grained attention for music generation.
Specifically, with the fine-grained attention, a token of a specific bar directly attends to all the tokens of the bars that are most relevant to music structures.
With the coarse-grained attention, a token only attends to the summarization of the other bars rather than each token of them so as to reduce the computational cost.
arXiv Detail & Related papers (2022-10-19T07:31:56Z) - Music Generation Using an LSTM [52.77024349608834]
Long Short-Term Memory (LSTM) network structures have proven to be very useful for making predictions for the next output in a series.
We demonstrate an approach of music generation using Recurrent Neural Networks (RNN)
We provide a brief synopsis of the intuition, theory, and application of LSTMs in music generation, develop and present the network we found to best achieve this goal, identify and address issues and challenges faced, and include potential future improvements for our network.
arXiv Detail & Related papers (2022-03-23T00:13:41Z) - Rhythm, Chord and Melody Generation for Lead Sheets using Recurrent
Neural Networks [5.57310999362848]
We propose a two-stage LSTM-based model for lead sheet generation, in which the harmonic and rhythmic templates of the song are produced first, after which, in a second stage, a sequence of melody notes is generated conditioned on these templates.
A subjective listening test shows that our approach outperforms the baselines and increases perceived musical coherence.
arXiv Detail & Related papers (2020-02-21T09:36:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.