A principled distributional approach to trajectory similarity
measurement
- URL: http://arxiv.org/abs/2301.00393v1
- Date: Sun, 1 Jan 2023 12:35:07 GMT
- Title: A principled distributional approach to trajectory similarity
measurement
- Authors: Yufan Wang, Kai Ming Ting, Yuanyi Shang
- Abstract summary: This paper proposes a powerful way to represent trajectories and measure the similarity between two trajectories using a distributional kernel.
A distributional kernel is used for the very first time for trajectory representation and similarity measurement.
We show the generality of this new approach in three applications: (a) trajectory anomaly detection, (b) anomalous sub-trajectory detection, and (c) trajectory pattern mining.
- Score: 8.316979146894989
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing measures and representations for trajectories have two longstanding
fundamental shortcomings, i.e., they are computationally expensive and they can
not guarantee the `uniqueness' property of a distance function: dist(X,Y) = 0
if and only if X=Y, where $X$ and $Y$ are two trajectories. This paper proposes
a simple yet powerful way to represent trajectories and measure the similarity
between two trajectories using a distributional kernel to address these
shortcomings. It is a principled approach based on kernel mean embedding which
has a strong theoretical underpinning. It has three distinctive features in
comparison with existing approaches. (1) A distributional kernel is used for
the very first time for trajectory representation and similarity measurement.
(2) It does not rely on point-to-point distances which are used in most
existing distances for trajectories. (3) It requires no learning, unlike
existing learning and deep learning approaches. We show the generality of this
new approach in three applications: (a) trajectory anomaly detection, (b)
anomalous sub-trajectory detection, and (c) trajectory pattern mining. We
identify that the distributional kernel has (i) a unique data-dependent
property and the above uniqueness property which are the key factors that lead
to its superior task-specific performance; and (ii) runtime orders of magnitude
faster than existing distance measures.
Related papers
- Kernel Trace Distance: Quantum Statistical Metric between Measures through RKHS Density Operators [11.899035547580201]
We introduce a novel distance between measures that compares them through a Schatten norm of their kernel covariance operators.<n>We show that this new distance is an integral probability metric that can be framed between a Maximum Mean Discrepancy (MMD) and a Wasserstein distance.
arXiv Detail & Related papers (2025-07-08T14:56:44Z) - Wasserstein-based Kernels for Clustering: Application to Power Distribution Graphs [0.0]
This work explores kernel methods and Wasserstein distance metrics to develop a computationally tractable clustering framework.
The framework is flexible enough to be applied in various domains, such as graph analysis and image processing.
A case study involving two datasets of 879 and 34,920 power distribution graphs demonstrates the framework's effectiveness and efficiency.
arXiv Detail & Related papers (2025-03-18T15:40:55Z) - Learning Temporal Distances: Contrastive Successor Features Can Provide a Metric Structure for Decision-Making [66.27188304203217]
Temporal distances lie at the heart of many algorithms for planning, control, and reinforcement learning.
Prior attempts to define such temporal distances in settings have been stymied by an important limitation.
We show how successor features learned by contrastive learning form a temporal distance that does satisfy the triangle inequality.
arXiv Detail & Related papers (2024-06-24T19:36:45Z) - A Distribution-Based Threshold for Determining Sentence Similarity [0.0]
We present a solution to a semantic textual similarity (STS) problem in which it is necessary to match two sentences containing, as the only distinguishing factor, highly specific information.
The solution revolves around the use of a neural network, based on the siamese architecture, to create the distributions of the distances between similar and dissimilar pairs of sentences.
arXiv Detail & Related papers (2023-11-28T10:42:35Z) - Distribution-Based Trajectory Clustering [14.781854651899705]
Trajectory clustering enables the discovery of common patterns in trajectory data.
The distance measures employed have two challenges: high computational cost and low fidelity.
We propose to use a recent Isolation Distributional Kernel (IDK) as the main tool to meet all three challenges.
arXiv Detail & Related papers (2023-10-08T11:28:34Z) - Energy-Based Sliced Wasserstein Distance [47.18652387199418]
A key component of the sliced Wasserstein (SW) distance is the slicing distribution.
We propose to design the slicing distribution as an energy-based distribution that is parameter-free.
We then derive a novel sliced Wasserstein metric, energy-based sliced Waserstein (EBSW) distance.
arXiv Detail & Related papers (2023-04-26T14:28:45Z) - Detecting Rotated Objects as Gaussian Distributions and Its 3-D
Generalization [81.29406957201458]
Existing detection methods commonly use a parameterized bounding box (BBox) to model and detect (horizontal) objects.
We argue that such a mechanism has fundamental limitations in building an effective regression loss for rotation detection.
We propose to model the rotated objects as Gaussian distributions.
We extend our approach from 2-D to 3-D with a tailored algorithm design to handle the heading estimation.
arXiv Detail & Related papers (2022-09-22T07:50:48Z) - GraphWalks: Efficient Shape Agnostic Geodesic Shortest Path Estimation [93.60478281489243]
We propose a learnable network to approximate geodesic paths on 3D surfaces.
The proposed method provides efficient approximations of the shortest paths and geodesic distances estimations.
arXiv Detail & Related papers (2022-05-30T16:22:53Z) - On the Benefits of Large Learning Rates for Kernel Methods [110.03020563291788]
We show that a phenomenon can be precisely characterized in the context of kernel methods.
We consider the minimization of a quadratic objective in a separable Hilbert space, and show that with early stopping, the choice of learning rate influences the spectral decomposition of the obtained solution.
arXiv Detail & Related papers (2022-02-28T13:01:04Z) - Kernel distance measures for time series, random fields and other
structured data [71.61147615789537]
kdiff is a novel kernel-based measure for estimating distances between instances of structured data.
It accounts for both self and cross similarities across the instances and is defined using a lower quantile of the distance distribution.
Some theoretical results are provided for separability conditions using kdiff as a distance measure for clustering and classification problems.
arXiv Detail & Related papers (2021-09-29T22:54:17Z) - Non-Asymptotic Performance Guarantees for Neural Estimation of
$\mathsf{f}$-Divergences [22.496696555768846]
Statistical distances quantify the dissimilarity between probability distributions.
A modern method for estimating such distances from data relies on parametrizing a variational form by a neural network (NN) and optimizing it.
This paper explores this tradeoff by means of non-asymptotic error bounds, focusing on three popular choices of SDs.
arXiv Detail & Related papers (2021-03-11T19:47:30Z) - Schoenberg-Rao distances: Entropy-based and geometry-aware statistical
Hilbert distances [12.729120803225065]
We study a class of statistical Hilbert distances that we term the Schoenberg-Rao distances.
We derive novel closed-form distances between mixtures of Gaussian distributions.
Our method constitutes a practical alternative to Wasserstein distances and we illustrate its efficiency on a broad range of machine learning tasks.
arXiv Detail & Related papers (2020-02-19T18:48:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.