Outlier Detection for Trajectories via Flow-embeddings
- URL: http://arxiv.org/abs/2111.13235v1
- Date: Thu, 25 Nov 2021 19:58:48 GMT
- Title: Outlier Detection for Trajectories via Flow-embeddings
- Authors: Florian Frantzen and Jean-Baptiste Seby and Michael T. Schaub
- Abstract summary: We propose a method to detect outliers in empirically observed trajectories on a discretized manifold modeled by a simplicial complex.
Our approach is similar to spectral embeddings such as diffusion-maps and Laplacian eigenmaps, that construct embeddings from the eigenvectors of the graph Laplacian associated with low eigenvalues.
We show how this technique can single out trajectories that behave (topologically) different compared to typical trajectories, and illustrate the performance of our approach with both synthetic and empirical data.
- Score: 2.66418345185993
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a method to detect outliers in empirically observed trajectories
on a discrete or discretized manifold modeled by a simplicial complex. Our
approach is similar to spectral embeddings such as diffusion-maps and Laplacian
eigenmaps, that construct vertex embeddings from the eigenvectors of the graph
Laplacian associated with low eigenvalues. Here we consider trajectories as
edge-flow vectors defined on a simplicial complex, a higher-order
generalization of graphs, and use the Hodge 1-Laplacian of the simplicial
complex to derive embeddings of these edge-flows. By projecting trajectory
vectors onto the eigenspace of the Hodge 1-Laplacian associated to small
eigenvalues, we can characterize the behavior of the trajectories relative to
the homology of the complex, which corresponds to holes in the underlying
space. This enables us to classify trajectories based on simply interpretable,
low-dimensional statistics. We show how this technique can single out
trajectories that behave (topologically) different compared to typical
trajectories, and illustrate the performance of our approach with both
synthetic and empirical data.
Related papers
- Lines of Thought in Large Language Models [3.281128493853064]
Large Language Models achieve next-token prediction by transporting a vectorized piece of text across an accompanying embedding space.
We aim to characterize the statistical properties of ensembles of these 'lines of thought'
We find it remarkable that the vast complexity of such large models can be reduced to a much simpler form, and we reflect on implications.
arXiv Detail & Related papers (2024-10-02T13:31:06Z) - Mixed Gaussian Flow for Diverse Trajectory Prediction [78.00204650749453]
We propose a flow-based model to transform a mixed Gaussian prior into the future trajectory manifold.
The model shows a better capacity for generating diverse trajectory patterns.
We also demonstrate that it can generate diverse, controllable, and out-of-distribution trajectories.
arXiv Detail & Related papers (2024-02-19T15:48:55Z) - Disentangling the Spectral Properties of the Hodge Laplacian: Not All Small Eigenvalues Are Equal [5.079602839359521]
The Hodge Laplacian has come into focus as a generalisation of the ordinary Laplacian for higher-order graph models such as simplicial and cellular complexes.
We introduce the notion of persistent eigenvector similarity and provide a method to track individual harmonic, curl, and gradient eigenvectors/-values.
We also use our insights to introduce a novel form of Hodge spectral clustering and to classify edges and higher-order simplices.
arXiv Detail & Related papers (2023-11-24T12:00:50Z) - Gradient-Based Feature Learning under Structured Data [57.76552698981579]
In the anisotropic setting, the commonly used spherical gradient dynamics may fail to recover the true direction.
We show that appropriate weight normalization that is reminiscent of batch normalization can alleviate this issue.
In particular, under the spiked model with a suitably large spike, the sample complexity of gradient-based training can be made independent of the information exponent.
arXiv Detail & Related papers (2023-09-07T16:55:50Z) - Convolutional Filtering on Sampled Manifolds [122.06927400759021]
We show that convolutional filtering on a sampled manifold converges to continuous manifold filtering.
Our findings are further demonstrated empirically on a problem of navigation control.
arXiv Detail & Related papers (2022-11-20T19:09:50Z) - Towards Modeling and Resolving Singular Parameter Spaces using
Stratifolds [18.60761407945024]
In learning dynamics, singularities can act as attractors on the learning trajectory and, therefore, negatively influence the convergence speed of models.
We propose a general approach to circumvent the problem arising from singularities by using stratifolds.
We empirically show that using (natural) gradient descent on the smooth manifold approximation instead of the singular space allows us to avoid the attractor behavior and therefore improve the convergence speed in learning.
arXiv Detail & Related papers (2021-12-07T14:42:45Z) - A Differential Geometry Perspective on Orthogonal Recurrent Models [56.09491978954866]
We employ tools and insights from differential geometry to offer a novel perspective on orthogonal RNNs.
We show that orthogonal RNNs may be viewed as optimizing in the space of divergence-free vector fields.
Motivated by this observation, we study a new recurrent model, which spans the entire space of vector fields.
arXiv Detail & Related papers (2021-02-18T19:39:22Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Manifold Learning via Manifold Deflation [105.7418091051558]
dimensionality reduction methods provide a valuable means to visualize and interpret high-dimensional data.
Many popular methods can fail dramatically, even on simple two-dimensional Manifolds.
This paper presents an embedding method for a novel, incremental tangent space estimator that incorporates global structure as coordinates.
Empirically, we show our algorithm recovers novel and interesting embeddings on real-world and synthetic datasets.
arXiv Detail & Related papers (2020-07-07T10:04:28Z) - The Boomerang Sampler [4.588028371034406]
This paper introduces the Boomerang Sampler as a novel class of continuous-time non-reversible Markov chain Monte Carlo algorithms.
We demonstrate that the method is easy to implement and demonstrate empirically that it can out-perform existing benchmark piecewise deterministic Markov processes.
arXiv Detail & Related papers (2020-06-24T14:52:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.