A Short Note on Analyzing Sequence Complexity in Trajectory Prediction
Benchmarks
- URL: http://arxiv.org/abs/2004.04677v2
- Date: Thu, 28 May 2020 15:04:54 GMT
- Title: A Short Note on Analyzing Sequence Complexity in Trajectory Prediction
Benchmarks
- Authors: Ronny Hug, Stefan Becker, Wolfgang H\"ubner and Michael Arens
- Abstract summary: An approach for determining a dataset representation in terms of a small set of distinguishable sub-sequences is proposed.
A first proof of concept on synthetically generated and real-world datasets shows the viability of the approach.
- Score: 8.870188183999852
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The analysis and quantification of sequence complexity is an open problem
frequently encountered when defining trajectory prediction benchmarks. In order
to enable a more informative assembly of a data basis, an approach for
determining a dataset representation in terms of a small set of distinguishable
prototypical sub-sequences is proposed. The approach employs a sequence
alignment followed by a learning vector quantization (LVQ) stage. A first proof
of concept on synthetically generated and real-world datasets shows the
viability of the approach.
Related papers
- Entrywise Inference for Missing Panel Data: A Simple and Instance-Optimal Approach [27.301741710016223]
We consider inferential questions associated with the missing data version of panel data induced by staggered adoption.
We develop and analyze a data-driven procedure for constructing entrywise confidence intervals with pre-specified coverage.
We prove non-asymptotic and high-probability bounds on its error in estimating each missing entry.
arXiv Detail & Related papers (2024-01-24T18:58:18Z) - Linear Partial Monitoring for Sequential Decision-Making: Algorithms,
Regret Bounds and Applications [70.67112733968654]
Partial monitoring is an expressive framework for sequential decision-making.
We present a simple and unified analysis of partial monitoring, and further extend the model to the contextual and kernelized setting.
arXiv Detail & Related papers (2023-02-07T18:58:25Z) - Estimating leverage scores via rank revealing methods and randomization [50.591267188664666]
We study algorithms for estimating the statistical leverage scores of rectangular dense or sparse matrices of arbitrary rank.
Our approach is based on combining rank revealing methods with compositions of dense and sparse randomized dimensionality reduction transforms.
arXiv Detail & Related papers (2021-05-23T19:21:55Z) - Parsimonious Inference [0.0]
Parsimonious inference is an information-theoretic formulation of inference over arbitrary architectures.
Our approaches combine efficient encodings with prudent sampling strategies to construct predictive ensembles without cross-validation.
arXiv Detail & Related papers (2021-03-03T04:13:14Z) - Revisiting the Sample Complexity of Sparse Spectrum Approximation of
Gaussian Processes [60.479499225746295]
We introduce a new scalable approximation for Gaussian processes with provable guarantees which hold simultaneously over its entire parameter space.
Our approximation is obtained from an improved sample complexity analysis for sparse spectrum Gaussian processes (SSGPs)
arXiv Detail & Related papers (2020-11-17T05:41:50Z) - Representation Learning for Sequence Data with Deep Autoencoding
Predictive Components [96.42805872177067]
We propose a self-supervised representation learning method for sequence data, based on the intuition that useful representations of sequence data should exhibit a simple structure in the latent space.
We encourage this latent structure by maximizing an estimate of predictive information of latent feature sequences, which is the mutual information between past and future windows at each time step.
We demonstrate that our method recovers the latent space of noisy dynamical systems, extracts predictive features for forecasting tasks, and improves automatic speech recognition when used to pretrain the encoder on large amounts of unlabeled data.
arXiv Detail & Related papers (2020-10-07T03:34:01Z) - Predicting Temporal Sets with Deep Neural Networks [50.53727580527024]
We propose an integrated solution based on the deep neural networks for temporal sets prediction.
A unique perspective is to learn element relationship by constructing set-level co-occurrence graph.
We design an attention-based module to adaptively learn the temporal dependency of elements and sets.
arXiv Detail & Related papers (2020-06-20T03:29:02Z) - Hierarchical regularization networks for sparsification based learning
on noisy datasets [0.0]
hierarchy follows from approximation spaces identified at successively finer scales.
For promoting model generalization at each scale, we also introduce a novel, projection based penalty operator across multiple dimension.
Results show the performance of the approach as a data reduction and modeling strategy on both synthetic and real datasets.
arXiv Detail & Related papers (2020-06-09T18:32:24Z) - Quantifying the Complexity of Standard Benchmarking Datasets for
Long-Term Human Trajectory Prediction [8.870188183999852]
We propose an approach for quantifying the amount of information contained in a dataset from a prototype-based dataset representation.
A large-scale complexity analysis is conducted on several human trajectory prediction benchmarking datasets.
arXiv Detail & Related papers (2020-05-28T12:00:41Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.