Understanding and Compressing Music with Maximal Transformable Patterns
- URL: http://arxiv.org/abs/2201.11085v1
- Date: Wed, 26 Jan 2022 17:47:26 GMT
- Title: Understanding and Compressing Music with Maximal Transformable Patterns
- Authors: David Meredith
- Abstract summary: We present an algorithm that discovers maximal patterns in a point set, $DinmathbbRk$.
We also present a second algorithm that discovers the set of occurrences for each of these maximal patterns.
We evaluate the new compression algorithm with three classes of differing complexity on the task of classifying folk-song melodies into tune families.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a polynomial-time algorithm that discovers all maximal patterns in
a point set, $D\in\mathbb{R}^k$, that are related by transformations in a
user-specified class, $F$, of bijections over $\mathbb{R}^k$. We also present a
second algorithm that discovers the set of occurrences for each of these
maximal patterns and then uses compact encodings of these occurrence sets to
compute a losslessly compressed encoding of the input point set. This encoding
takes the form of a set of pairs, $E=\left\lbrace\left\langle P_1,
T_1\right\rangle,\left\langle P_2, T_2\right\rangle,\ldots\left\langle
P_{\ell}, T_{\ell}\right\rangle\right\rbrace$, where each $\langle
P_i,T_i\rangle$ consists of a maximal pattern, $P_i\subseteq D$, and a set,
$T_i\subset F$, of transformations that map $P_i$ onto other subsets of $D$.
Each transformation is encoded by a vector of real values that uniquely
identifies it within $F$ and the length of this vector is used as a measure of
the complexity of $F$. We evaluate the new compression algorithm with three
transformation classes of differing complexity, on the task of classifying
folk-song melodies into tune families. The most complex of the classes tested
includes all combinations of the musical transformations of transposition,
inversion, retrograde, augmentation and diminution. We found that broadening
the transformation class improved performance on this task. However, it did
not, on average, improve compression factor, which may be due to the datasets
(in this case, folk-song melodies) being too short and simple to benefit from
the potentially greater number of pattern relationships that are discoverable
with larger transformation classes.
Related papers
- Projection by Convolution: Optimal Sample Complexity for Reinforcement Learning in Continuous-Space MDPs [56.237917407785545]
We consider the problem of learning an $varepsilon$-optimal policy in a general class of continuous-space Markov decision processes (MDPs) having smooth Bellman operators.
Key to our solution is a novel projection technique based on ideas from harmonic analysis.
Our result bridges the gap between two popular but conflicting perspectives on continuous-space MDPs.
arXiv Detail & Related papers (2024-05-10T09:58:47Z) - Chain of Thought Empowers Transformers to Solve Inherently Serial Problems [57.58801785642868]
Chain of thought (CoT) is a highly effective method to improve the accuracy of large language models (LLMs) on arithmetics and symbolic reasoning tasks.
This work provides a theoretical understanding of the power of CoT for decoder-only transformers through the lens of expressiveness.
arXiv Detail & Related papers (2024-02-20T10:11:03Z) - Learning to Understand: Identifying Interactions via the Möbius Transform [18.987216240237483]
We use the M"obius transform to find interpretable representations of learned functions.
A robust version of this algorithm withstands noise and maintains this complexity.
In several examples, we observe that representations generated via the M"obius transform are up to twice as faithful to the original function.
arXiv Detail & Related papers (2024-02-04T22:47:34Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - A Formal Perspective on Byte-Pair Encoding [100.75374173565548]
Byte-Pair imation (BPE) is a popular algorithm used for tokenizing data in NLP, despite being devised initially as a compression method.
We provide a faster implementation of BPE which improves the runtime complexity from $mathcalOleft(N Mright)$ to $mathcalOleft(N log Mright)$, where $N$ is the sequence length and $M$ is the merge count.
arXiv Detail & Related papers (2023-06-29T10:29:23Z) - Empirical complexity of comparator-based nearest neighbor descent [0.0]
A Java parallel streams implementation of the $K$-nearest neighbor descent algorithm is presented.
Experiments with the Kullback-Leibler divergence Comparator support the prediction that the number of rounds of $K$-nearest neighbor updates need not exceed twice the diameter.
arXiv Detail & Related papers (2022-01-30T21:37:53Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - Faster Binary Embeddings for Preserving Euclidean Distances [9.340611077939828]
We propose a fast, distance-preserving, binary embedding algorithm to transform a dataset $mathcalTsubseteqmathbbRn$ into binary sequences in the cube $pm 1m$.
Our method is both fast and memory efficient, with time complexity $O(m)$ and space complexity $O(m)$.
arXiv Detail & Related papers (2020-10-01T22:41:41Z) - On Efficient Low Distortion Ultrametric Embedding [18.227854382422112]
A widely-used method to preserve the underlying hierarchical structure of the data is to find an embedding of the data into a tree or an ultrametric.
In this paper, we provide a new algorithm which takes as input a set of points isometric in $mathbbRd2 (for universal constant $rho>1$) to output an ultrametric $Delta.
We show that the output of our algorithm is comparable to the output of the linkage algorithms while achieving a much faster running time.
arXiv Detail & Related papers (2020-08-15T11:06:45Z) - Model-Free Reinforcement Learning: from Clipped Pseudo-Regret to Sample
Complexity [59.34067736545355]
Given an MDP with $S$ states, $A$ actions, the discount factor $gamma in (0,1)$, and an approximation threshold $epsilon > 0$, we provide a model-free algorithm to learn an $epsilon$-optimal policy.
For small enough $epsilon$, we show an improved algorithm with sample complexity.
arXiv Detail & Related papers (2020-06-06T13:34:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.