LoopNet: Musical Loop Synthesis Conditioned On Intuitive Musical
Parameters
- URL: http://arxiv.org/abs/2105.10371v1
- Date: Fri, 21 May 2021 14:24:34 GMT
- Title: LoopNet: Musical Loop Synthesis Conditioned On Intuitive Musical
Parameters
- Authors: Pritish Chandna, Ant\'onio Ramires, Xavier Serra, Emilia G\'omez
- Abstract summary: LoopNet is a feed-forward generative model for creating loops conditioned on intuitive parameters.
We leverage Music Information Retrieval (MIR) models as well as a large collection of public loop samples in our study.
- Score: 12.72202888016628
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Loops, seamlessly repeatable musical segments, are a cornerstone of modern
music production. Contemporary artists often mix and match various sampled or
pre-recorded loops based on musical criteria such as rhythm, harmony and
timbral texture to create compositions. Taking such criteria into account, we
present LoopNet, a feed-forward generative model for creating loops conditioned
on intuitive parameters. We leverage Music Information Retrieval (MIR) models
as well as a large collection of public loop samples in our study and use the
Wave-U-Net architecture to map control parameters to audio. We also evaluate
the quality of the generated audio and propose intuitive controls for composers
to map the ideas in their minds to an audio loop.
Related papers
- MusicFlow: Cascaded Flow Matching for Text Guided Music Generation [53.63948108922333]
MusicFlow is a cascaded text-to-music generation model based on flow matching.
We leverage masked prediction as the training objective, enabling the model to generalize to other tasks such as music infilling and continuation.
arXiv Detail & Related papers (2024-10-27T15:35:41Z) - MuVi: Video-to-Music Generation with Semantic Alignment and Rhythmic Synchronization [52.498942604622165]
This paper presents MuVi, a framework to generate music that aligns with video content.
MuVi analyzes video content through a specially designed visual adaptor to extract contextually and temporally relevant features.
We show that MuVi demonstrates superior performance in both audio quality and temporal synchronization.
arXiv Detail & Related papers (2024-10-16T18:44:56Z) - MusiConGen: Rhythm and Chord Control for Transformer-Based Text-to-Music Generation [19.878013881045817]
MusiConGen is a temporally-conditioned Transformer-based text-to-music model.
It integrates automatically-extracted rhythm and chords as the condition signal.
We show that MusiConGen can generate realistic backing track music that aligns well with the specified conditions.
arXiv Detail & Related papers (2024-07-21T05:27:53Z) - Simple and Controllable Music Generation [94.61958781346176]
MusicGen is a single Language Model (LM) that operates over several streams of compressed discrete music representation, i.e., tokens.
Unlike prior work, MusicGen is comprised of a single-stage transformer LM together with efficient token interleaving patterns.
arXiv Detail & Related papers (2023-06-08T15:31:05Z) - GETMusic: Generating Any Music Tracks with a Unified Representation and
Diffusion Framework [58.64512825534638]
Symbolic music generation aims to create musical notes, which can help users compose music.
We introduce a framework known as GETMusic, with GET'' standing for GEnerate music Tracks''
GETScore represents musical notes as tokens and organizes tokens in a 2D structure, with tracks stacked vertically and progressing horizontally over time.
Our proposed representation, coupled with the non-autoregressive generative model, empowers GETMusic to generate music with any arbitrary source-target track combinations.
arXiv Detail & Related papers (2023-05-18T09:53:23Z) - Long-Term Rhythmic Video Soundtracker [37.082768654951465]
We present Long-Term Rhythmic Video Soundtracker (LORIS), a novel framework to synthesize long-term conditional waveforms.
We extend our model's applicability from dances to multiple sports scenarios such as floor exercise and figure skating.
Our model generates long-term soundtracks with state-of-the-art musical quality and rhythmic correspondence.
arXiv Detail & Related papers (2023-05-02T10:58:29Z) - Quantized GAN for Complex Music Generation from Dance Videos [48.196705493763986]
We present Dance2Music-GAN (D2M-GAN), a novel adversarial multi-modal framework that generates musical samples conditioned on dance videos.
Our proposed framework takes dance video frames and human body motion as input, and learns to generate music samples that plausibly accompany the corresponding input.
arXiv Detail & Related papers (2022-04-01T17:53:39Z) - CycleDRUMS: Automatic Drum Arrangement For Bass Lines Using CycleGAN [12.93891163150604]
CycleDRUMS is a novel method for generating drums given a bass line.
After converting the waveform of the bass into a mel-spectrogram, we are able to automatically generate original drums that follow the beat.
arXiv Detail & Related papers (2021-04-01T09:17:48Z) - A framework to compare music generative models using automatic
evaluation metrics extended to rhythm [69.2737664640826]
This paper takes the framework proposed in a previous research that did not consider rhythm to make a series of design decisions, then, rhythm support is added to evaluate the performance of two RNN memory cells in the creation of monophonic music.
The model considers the handling of music transposition and the framework evaluates the quality of the generated pieces using automatic quantitative metrics based on geometry which have rhythm support added as well.
arXiv Detail & Related papers (2021-01-19T15:04:46Z) - Music SketchNet: Controllable Music Generation via Factorized
Representations of Pitch and Rhythm [42.694266687511906]
Music SketchNet is a neural network framework that allows users to specify partial musical ideas guiding automatic music generation.
We focus on generating the missing measures in incomplete monophonic musical pieces, conditioned on surrounding context.
We demonstrate that our model can successfully incorporate user-specified snippets during the generation process.
arXiv Detail & Related papers (2020-08-04T02:49:57Z) - Continuous Melody Generation via Disentangled Short-Term Representations
and Structural Conditions [14.786601824794369]
We present a model for composing melodies given a user specified symbolic scenario combined with a previous music context.
Our model is capable of generating long melodies by regarding 8-beat note sequences as basic units, and shares consistent rhythm pattern structure with another specific song.
Results show that the music generated by our model tends to have salient repetition structures, rich motives, and stable rhythm patterns.
arXiv Detail & Related papers (2020-02-05T06:23:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.