GuitarFlow: Realistic Electric Guitar Synthesis From Tablatures via Flow Matching and Style Transfer
- URL: http://arxiv.org/abs/2510.21872v1
- Date: Thu, 23 Oct 2025 13:31:41 GMT
- Title: GuitarFlow: Realistic Electric Guitar Synthesis From Tablatures via Flow Matching and Style Transfer
- Authors: Jackson Loth, Pedro Sarmento, Mark Sandler, Mathieu Barthet,
- Abstract summary: We introduce GuitarFlow, a model designed specifically for electric guitar synthesis.<n>The generative process is guided using tablatures, an ubiquitous and intuitive guitar-specific symbolic format.<n>We show significant improvement in the realism of the generated guitar audio from tablatures.
- Score: 7.72498447842112
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Music generation in the audio domain using artificial intelligence (AI) has witnessed steady progress in recent years. However for some instruments, particularly the guitar, controllable instrument synthesis remains limited in expressivity. We introduce GuitarFlow, a model designed specifically for electric guitar synthesis. The generative process is guided using tablatures, an ubiquitous and intuitive guitar-specific symbolic format. The tablature format easily represents guitar-specific playing techniques (e.g. bends, muted strings and legatos), which are more difficult to represent in other common music notation formats such as MIDI. Our model relies on an intermediary step of first rendering the tablature to audio using a simple sample-based virtual instrument, then performing style transfer using Flow Matching in order to transform the virtual instrument audio into more realistic sounding examples. This results in a model that is quick to train and to perform inference, requiring less than 6 hours of training data. We present the results of objective evaluation metrics, together with a listening test, in which we show significant improvement in the realism of the generated guitar audio from tablatures.
Related papers
- A Machine Learning Approach for MIDI to Guitar Tablature Conversion [21.416973100105633]
This paper presents a method for assigning guitar tablature notation to a given MIDI-based musical part.<n>The strategy is based on machine learning and requires a basic assumption about how much fingers can stretch on a fretboard.<n>The proposed method also examines the transcription of music pieces that was not meant to be played or could not possibly be played by a guitar.
arXiv Detail & Related papers (2025-10-12T14:01:01Z) - Transcribing Rhythmic Patterns of the Guitar Track in Polyphonic Music [46.69593319852797]
We transcribe the rhythmic patterns in 410 popular songs and record cover versions where the guitar tracks followed those transcriptions.<n>We detect individual strums within the separated guitar audio, using a pre-trained foundation model (MERT) as a backbone.<n>We show that it is possible to transcribe the rhythmic patterns of the guitar track in polyphonic music with quite high accuracy.
arXiv Detail & Related papers (2025-10-07T10:22:31Z) - SCORE-SET: A dataset of GuitarPro files for Music Phrase Generation and Sequence Learning [0.0]
The dataset is derived from MIDI notes in MAESTRO and GiantMIDI which have been adapted into rhythm guitar tracks.<n>These tracks are processed to include a variety of expression settings typical of guitar performance, such as bends, slides, vibrato, and palm muting.
arXiv Detail & Related papers (2025-07-24T18:13:12Z) - Scaling Self-Supervised Representation Learning for Symbolic Piano Performance [52.661197827466886]
We study the capabilities of generative autoregressive transformer models trained on large amounts of symbolic solo-piano transcriptions.<n>We use a comparatively smaller, high-quality subset to finetune models to produce musical continuations, perform symbolic classification tasks, and produce general-purpose contrastive MIDI embeddings.
arXiv Detail & Related papers (2025-06-30T14:00:14Z) - MIDI-to-Tab: Guitar Tablature Inference via Masked Language Modeling [6.150307957212576]
We introduce a novel deep learning solution to symbolic guitar tablature estimation.
We train an encoder-decoder Transformer model in a masked language modeling paradigm to assign notes to strings.
The model is first pre-trained on DadaGP, a dataset of over 25K tablatures, and then fine-tuned on a curated set of professionally transcribed guitar performances.
arXiv Detail & Related papers (2024-08-09T12:25:23Z) - Expressive Acoustic Guitar Sound Synthesis with an Instrument-Specific
Input Representation and Diffusion Outpainting [9.812666469580872]
We propose an expressive acoustic guitar sound synthesis model with a customized input representation to the instrument.
We implement the proposed approach using diffusion-based outpainting which can generate audio with long-term consistency.
Our proposed model has higher audio quality than the baseline model and generates more realistic timbre sounds.
arXiv Detail & Related papers (2024-01-24T14:44:01Z) - Modeling Bends in Popular Music Guitar Tablatures [49.64902130083662]
Tablature notation is widely used in popular music to transcribe and share guitar musical content.
This paper focuses on bends, which enable to progressively shift the pitch of a note, therefore circumventing physical limitations of the discrete fretted fingerboard.
Experiments are performed on a corpus of 932 lead guitar tablatures of popular music and show that a decision tree successfully predicts bend occurrences with an F1 score of 0.71 anda limited amount of false positive predictions.
arXiv Detail & Related papers (2023-08-22T07:50:58Z) - Simple and Controllable Music Generation [94.61958781346176]
MusicGen is a single Language Model (LM) that operates over several streams of compressed discrete music representation, i.e., tokens.
Unlike prior work, MusicGen is comprised of a single-stage transformer LM together with efficient token interleaving patterns.
arXiv Detail & Related papers (2023-06-08T15:31:05Z) - GTR-CTRL: Instrument and Genre Conditioning for Guitar-Focused Music
Generation with Transformers [14.025337055088102]
We use the DadaGP dataset for guitar tab music generation, a corpus of over 26k songs in GuitarPro and token formats.
We introduce methods to condition a Transformer-XL deep learning model to generate guitar tabs based on desired instrumentation and genre.
Results indicate that the GTR-CTRL methods provide more flexibility and control for guitar-focused symbolic music generation than an unconditioned model.
arXiv Detail & Related papers (2023-02-10T17:43:03Z) - Strumming to the Beat: Audio-Conditioned Contrastive Video Textures [112.6140796961121]
We introduce a non-parametric approach for infinite video texture synthesis using a representation learned via contrastive learning.
We take inspiration from Video Textures, which showed that plausible new videos could be generated from a single one by stitching its frames together in a novel yet consistent order.
Our model outperforms baselines on human perceptual scores, can handle a diverse range of input videos, and can combine semantic and audio-visual cues in order to synthesize videos that synchronize well with an audio signal.
arXiv Detail & Related papers (2021-04-06T17:24:57Z) - Music Gesture for Visual Sound Separation [121.36275456396075]
"Music Gesture" is a keypoint-based structured representation to explicitly model the body and finger movements of musicians when they perform music.
We first adopt a context-aware graph network to integrate visual semantic context with body dynamics, and then apply an audio-visual fusion model to associate body movements with the corresponding audio signals.
arXiv Detail & Related papers (2020-04-20T17:53:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.