Deep Declarative Dynamic Time Warping for End-to-End Learning of
Alignment Paths
- URL: http://arxiv.org/abs/2303.10778v1
- Date: Sun, 19 Mar 2023 21:58:37 GMT
- Title: Deep Declarative Dynamic Time Warping for End-to-End Learning of
Alignment Paths
- Authors: Ming Xu and Sourav Garg and Michael Milford and Stephen Gould
- Abstract summary: This paper addresses learning end-to-end models for time series data that include a temporal alignment step via dynamic time warping (DTW)
We propose a DTW layer based around bi-level optimisation and deep declarative networks, which we name DecDTW.
We show that this property is particularly useful for applications where downstream loss functions are defined on the optimal alignment path itself.
- Score: 54.53208538517505
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper addresses learning end-to-end models for time series data that
include a temporal alignment step via dynamic time warping (DTW). Existing
approaches to differentiable DTW either differentiate through a fixed warping
path or apply a differentiable relaxation to the min operator found in the
recursive steps used to solve the DTW problem. We instead propose a DTW layer
based around bi-level optimisation and deep declarative networks, which we name
DecDTW. By formulating DTW as a continuous, inequality constrained optimisation
problem, we can compute gradients for the solution of the optimal alignment
(with respect to the underlying time series) using implicit differentiation. An
interesting byproduct of this formulation is that DecDTW outputs the optimal
warping path between two time series as opposed to a soft approximation,
recoverable from Soft-DTW. We show that this property is particularly useful
for applications where downstream loss functions are defined on the optimal
alignment path itself. This naturally occurs, for instance, when learning to
improve the accuracy of predicted alignments against ground truth alignments.
We evaluate DecDTW on two such applications, namely the audio-to-score
alignment task in music information retrieval and the visual place recognition
task in robotics, demonstrating state-of-the-art results in both.
Related papers
- OTW: Optimal Transport Warping for Time Series [75.69837166816501]
Dynamic Time Warping (DTW) has become the pragmatic choice for measuring distance between time series.
It suffers from unavoidable quadratic time complexity when the optimal alignment matrix needs to be computed exactly.
We introduce a new metric for time series data based on the Optimal Transport framework, called Optimal Transport Warping (OTW)
arXiv Detail & Related papers (2023-06-01T12:45:00Z) - Approximating DTW with a convolutional neural network on EEG data [9.409281517596396]
We propose a fast and differentiable approximation of Dynamic Time Wrapping (DTW)
We show that our methods achieve at least the same level of accuracy as other DTW main approximations with higher computational efficiency.
arXiv Detail & Related papers (2023-01-30T13:27:47Z) - Improved Convergence Rate of Stochastic Gradient Langevin Dynamics with
Variance Reduction and its Application to Optimization [50.83356836818667]
gradient Langevin Dynamics is one of the most fundamental algorithms to solve non-eps optimization problems.
In this paper, we show two variants of this kind, namely the Variance Reduced Langevin Dynamics and the Recursive Gradient Langevin Dynamics.
arXiv Detail & Related papers (2022-03-30T11:39:00Z) - A Two-Time-Scale Stochastic Optimization Framework with Applications in Control and Reinforcement Learning [13.908826484332282]
We study a new two-time-scale gradient method for solving optimization problems.
Our first contribution is to characterize the finite-time complexity of the proposed two-time-scale gradient algorithm.
We apply our framework to gradient-based policy evaluation algorithms in reinforcement learning.
arXiv Detail & Related papers (2021-09-29T23:15:23Z) - Nesterov Accelerated ADMM for Fast Diffeomorphic Image Registration [63.15453821022452]
Recent developments in approaches based on deep learning have achieved sub-second runtimes for DiffIR.
We propose a simple iterative scheme that functionally composes intermediate non-stationary velocity fields.
We then propose a convex optimisation model that uses a regularisation term of arbitrary order to impose smoothness on these velocity fields.
arXiv Detail & Related papers (2021-09-26T19:56:45Z) - Drop-DTW: Aligning Common Signal Between Sequences While Dropping
Outliers [33.174893836302005]
We introduce Drop-DTW, a novel algorithm that aligns the common signal between the sequences while automatically dropping the outlier elements from the matching.
In our experiments, we show that Drop-DTW is a robust similarity measure for sequence retrieval and demonstrate its effectiveness as a training loss on diverse applications.
arXiv Detail & Related papers (2021-08-26T18:52:35Z) - TC-DTW: Accelerating Multivariate Dynamic Time Warping Through Triangle
Inequality and Point Clustering [6.502892821109196]
The most popular algorithm used today is still the one developed seventeen years ago.
The new solution, named TC-DTW, introduces Triangle Inequality and Point Clustering into the algorithm design.
In experiments on DTW-based nearest neighbor finding, the new solution avoids as much as 98% (60% average) DTW distance calculations and yields as much as 25X (7.5X average) speedups.
arXiv Detail & Related papers (2021-01-15T16:38:28Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.