Music Generation Using an LSTM
- URL: http://arxiv.org/abs/2203.12105v1
- Date: Wed, 23 Mar 2022 00:13:41 GMT
- Title: Music Generation Using an LSTM
- Authors: Michael Conner, Lucas Gral, Kevin Adams, David Hunger, Reagan Strelow,
and Alexander Neuwirth
- Abstract summary: Long Short-Term Memory (LSTM) network structures have proven to be very useful for making predictions for the next output in a series.
We demonstrate an approach of music generation using Recurrent Neural Networks (RNN)
We provide a brief synopsis of the intuition, theory, and application of LSTMs in music generation, develop and present the network we found to best achieve this goal, identify and address issues and challenges faced, and include potential future improvements for our network.
- Score: 52.77024349608834
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Over the past several years, deep learning for sequence modeling has grown in
popularity. To achieve this goal, LSTM network structures have proven to be
very useful for making predictions for the next output in a series. For
instance, a smartphone predicting the next word of a text message could use an
LSTM. We sought to demonstrate an approach of music generation using Recurrent
Neural Networks (RNN). More specifically, a Long Short-Term Memory (LSTM)
neural network. Generating music is a notoriously complicated task, whether
handmade or generated, as there are a myriad of components involved. Taking
this into account, we provide a brief synopsis of the intuition, theory, and
application of LSTMs in music generation, develop and present the network we
found to best achieve this goal, identify and address issues and challenges
faced, and include potential future improvements for our network.
Related papers
- Music Emotion Prediction Using Recurrent Neural Networks [8.867897390286815]
This study aims to enhance music recommendation systems and support therapeutic interventions by tailoring music to fit listeners' emotional states.
We utilize Russell's Emotion Quadrant to categorize music into four distinct emotional regions and develop models capable of accurately predicting these categories.
Our approach involves extracting a comprehensive set of audio features using Librosa and applying various recurrent neural network architectures, including standard RNNs, Bidirectional RNNs, and Long Short-Term Memory (LSTM) networks.
arXiv Detail & Related papers (2024-05-10T18:03:20Z) - MuPT: A Generative Symbolic Music Pretrained Transformer [56.09299510129221]
We explore the application of Large Language Models (LLMs) to the pre-training of music.
To address the challenges associated with misaligned measures from different tracks during generation, we propose a Synchronized Multi-Track ABC Notation (SMT-ABC Notation)
Our contributions include a series of models capable of handling up to 8192 tokens, covering 90% of the symbolic music data in our training set.
arXiv Detail & Related papers (2024-04-09T15:35:52Z) - Recurrent Neural Networks and Long Short-Term Memory Networks: Tutorial
and Survey [9.092591746522483]
This tutorial paper is on Recurrent Neural Network (RNN), Long Short-Term Memory Network (LSTM), and their variants.
We start with a dynamical system and backpropagation through time for RNN.
We discuss the problems of gradient vanishing and explosion in long-term dependencies.
Then, we introduce LSTM gates and cells, history and variants of LSTM, and Gated Recurrent Units (GRU)
arXiv Detail & Related papers (2023-04-22T18:22:10Z) - Sign Language Recognition via Skeleton-Aware Multi-Model Ensemble [71.97020373520922]
Sign language is commonly used by deaf or mute people to communicate.
We propose a novel Multi-modal Framework with a Global Ensemble Model (GEM) for isolated Sign Language Recognition ( SLR)
Our proposed SAM- SLR-v2 framework is exceedingly effective and achieves state-of-the-art performance with significant margins.
arXiv Detail & Related papers (2021-10-12T16:57:18Z) - Working Memory Connections for LSTM [51.742526187978726]
We show that Working Memory Connections constantly improve the performance of LSTMs on a variety of tasks.
Numerical results suggest that the cell state contains useful information that is worth including in the gate structure.
arXiv Detail & Related papers (2021-08-31T18:01:30Z) - Music Generation using Deep Learning [10.155748914174003]
The proposed approach takes ABC notations from the Nottingham dataset and encodes it to beefed as input for the neural networks.
The primary objective is to input the neural networks with an arbitrary note, let the network process and augment a sequence based on the note until a good piece of music is produced.
arXiv Detail & Related papers (2021-05-19T10:27:58Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - GANs & Reels: Creating Irish Music using a Generative Adversarial
Network [2.6604997762611204]
We present a method for algorithmic melody generation using a generative adversarial network without recurrent components.
Music generation has been successfully done using recurrent neural networks, where the model learns sequence information that can help create authentic sounding melodies.
arXiv Detail & Related papers (2020-10-29T17:16:22Z) - Object Tracking through Residual and Dense LSTMs [67.98948222599849]
Deep learning-based trackers based on LSTMs (Long Short-Term Memory) recurrent neural networks have emerged as a powerful alternative.
DenseLSTMs outperform Residual and regular LSTM, and offer a higher resilience to nuisances.
Our case study supports the adoption of residual-based RNNs for enhancing the robustness of other trackers.
arXiv Detail & Related papers (2020-06-22T08:20:17Z) - Attentional networks for music generation [5.012960295592238]
We propose a deep learning based music generation method in order to produce old style music particularly JAZZ with rehashed melodic structures.
Owing to the success in modelling long-term temporal dependencies in sequential data and its success in case of videos, Bi-LSTMs with attention serve as the natural choice and early utilization in music generation.
We validate in our experiments that Bi-LSTMs with attention are able to preserve the richness and technical nuances of the music performed.
arXiv Detail & Related papers (2020-02-06T13:26:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.