Vector Quantized Contrastive Predictive Coding for Template-based Music
Generation
- URL: http://arxiv.org/abs/2004.10120v1
- Date: Tue, 21 Apr 2020 15:58:17 GMT
- Title: Vector Quantized Contrastive Predictive Coding for Template-based Music
Generation
- Authors: Ga\"etan Hadjeres and L\'eopold Crestel
- Abstract summary: We propose a flexible method for generating variations of discrete sequences in which tokens can be grouped into basic units.
We show how these compressed representations can be used to generate variations of a template sequence by using an appropriate attention pattern in the Transformer architecture.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we propose a flexible method for generating variations of
discrete sequences in which tokens can be grouped into basic units, like
sentences in a text or bars in music. More precisely, given a template
sequence, we aim at producing novel sequences sharing perceptible similarities
with the original template without relying on any annotation; so our problem of
generating variations is intimately linked to the problem of learning relevant
high-level representations without supervision. Our contribution is two-fold:
First, we propose a self-supervised encoding technique, named Vector Quantized
Contrastive Predictive Coding which allows to learn a meaningful assignment of
the basic units over a discrete set of codes, together with mechanisms allowing
to control the information content of these learnt discrete representations.
Secondly, we show how these compressed representations can be used to generate
variations of a template sequence by using an appropriate attention pattern in
the Transformer architecture. We illustrate our approach on the corpus of J.S.
Bach chorales where we discuss the musical meaning of the learnt discrete codes
and show that our proposed method allows to generate coherent and high-quality
variations of a given template.
Related papers
- σ-GPTs: A New Approach to Autoregressive Models [19.84252724050016]
We show that by simply adding a positional encoding for the output, this order can be modulated on-the-fly per-sample.
We evaluate our method across various domains, including language modeling, path-solving, and aircraft vertical rate prediction.
arXiv Detail & Related papers (2024-04-15T08:22:47Z) - Disentanglement via Latent Quantization [60.37109712033694]
In this work, we construct an inductive bias towards encoding to and decoding from an organized latent space.
We demonstrate the broad applicability of this approach by adding it to both basic data-re (vanilla autoencoder) and latent-reconstructing (InfoGAN) generative models.
arXiv Detail & Related papers (2023-05-28T06:30:29Z) - Attributable and Scalable Opinion Summarization [79.87892048285819]
We generate abstractive summaries by decoding frequent encodings, and extractive summaries by selecting the sentences assigned to the same frequent encodings.
Our method is attributable, because the model identifies sentences used to generate the summary as part of the summarization process.
It scales easily to many hundreds of input reviews, because aggregation is performed in the latent space rather than over long sequences of tokens.
arXiv Detail & Related papers (2023-05-19T11:30:37Z) - A Contextual Latent Space Model: Subsequence Modulation in Melodic
Sequence [0.0]
Some generative models for sequences such as music and text allow us to edit only subsequences, given surrounding context sequences.
We propose a contextual latent space model (M) in order for users to be able to explore subsequence generation with a sense of direction in the generation space.
A context-informed prior and decoder constitute the generative model of CLSM, and a context position-informed is the inference model.
arXiv Detail & Related papers (2021-11-23T07:51:39Z) - Style Equalization: Unsupervised Learning of Controllable Generative
Sequence Models [23.649790871960644]
We tackle the training-inference mismatch encountered during unsupervised learning of controllable generative sequence models.
By introducing a style transformation module that we call style equalization, we enable training using different content and style samples.
Our models achieve state-of-the-art style replication with a similar mean style opinion score as the real data.
arXiv Detail & Related papers (2021-10-06T16:17:57Z) - Structured Reordering for Modeling Latent Alignments in Sequence
Transduction [86.94309120789396]
We present an efficient dynamic programming algorithm performing exact marginal inference of separable permutations.
The resulting seq2seq model exhibits better systematic generalization than standard models on synthetic problems and NLP tasks.
arXiv Detail & Related papers (2021-06-06T21:53:54Z) - Cross-Thought for Sentence Encoder Pre-training [89.32270059777025]
Cross-Thought is a novel approach to pre-training sequence encoder.
We train a Transformer-based sequence encoder over a large set of short sequences.
Experiments on question answering and textual entailment tasks demonstrate that our pre-trained encoder can outperform state-of-the-art encoders.
arXiv Detail & Related papers (2020-10-07T21:02:41Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.