Efficient Wind Speed Nowcasting with GPU-Accelerated Nearest Neighbors
Algorithm
- URL: http://arxiv.org/abs/2112.10408v1
- Date: Mon, 20 Dec 2021 09:15:27 GMT
- Title: Efficient Wind Speed Nowcasting with GPU-Accelerated Nearest Neighbors
Algorithm
- Authors: Arnaud Pannatier, Ricardo Picatoste, Fran\c{c}ois Fleuret
- Abstract summary: This paper proposes a simple yet efficient high-altitude wind nowcasting pipeline.
It processes efficiently a vast amount of live data recorded by airplanes over the whole airspace.
It creates a unique context for each point in the dataset and then extrapolates from it.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a simple yet efficient high-altitude wind nowcasting
pipeline. It processes efficiently a vast amount of live data recorded by
airplanes over the whole airspace and reconstructs the wind field with good
accuracy. It creates a unique context for each point in the dataset and then
extrapolates from it. As creating such context is computationally intensive,
this paper proposes a novel algorithm that reduces the time and memory cost by
efficiently fetching nearest neighbors in a data set whose elements are
organized along smooth trajectories that can be approximated with piece-wise
linear structures.
We introduce an efficient and exact strategy implemented through algebraic
tensorial operations, which is well-suited to modern GPU-based computing
infrastructure. This method employs a scalable Euclidean metric and allows
masking data points along one dimension. When applied, this method is more
efficient than plain Euclidean k-NN and other well-known data selection methods
such as KDTrees and provides a several-fold speedup. We provide an
implementation in PyTorch and a novel data set to allow the replication of
empirical results.
Related papers
- Scalable Data Assimilation with Message Passing [9.55393191483615]
We exploit the formulation of data assimilation as a Bayesian inference problem and apply a message-passing algorithm to solve the spatial inference problem.
We can scale the algorithm to very large grid sizes while retaining good accuracy and compute and memory requirements.
arXiv Detail & Related papers (2024-04-19T15:54:15Z) - Randomized Polar Codes for Anytime Distributed Machine Learning [66.46612460837147]
We present a novel distributed computing framework that is robust to slow compute nodes, and is capable of both approximate and exact computation of linear operations.
We propose a sequential decoding algorithm designed to handle real valued data while maintaining low computational complexity for recovery.
We demonstrate the potential applications of this framework in various contexts, such as large-scale matrix multiplication and black-box optimization.
arXiv Detail & Related papers (2023-09-01T18:02:04Z) - Linearized Wasserstein dimensionality reduction with approximation
guarantees [65.16758672591365]
LOT Wassmap is a computationally feasible algorithm to uncover low-dimensional structures in the Wasserstein space.
We show that LOT Wassmap attains correct embeddings and that the quality improves with increased sample size.
We also show how LOT Wassmap significantly reduces the computational cost when compared to algorithms that depend on pairwise distance computations.
arXiv Detail & Related papers (2023-02-14T22:12:16Z) - Efficient Dataset Distillation Using Random Feature Approximation [109.07737733329019]
We propose a novel algorithm that uses a random feature approximation (RFA) of the Neural Network Gaussian Process (NNGP) kernel.
Our algorithm provides at least a 100-fold speedup over KIP and can run on a single GPU.
Our new method, termed an RFA Distillation (RFAD), performs competitively with KIP and other dataset condensation algorithms in accuracy over a range of large-scale datasets.
arXiv Detail & Related papers (2022-10-21T15:56:13Z) - Communication-Efficient Adam-Type Algorithms for Distributed Data Mining [93.50424502011626]
We propose a class of novel distributed Adam-type algorithms (emphi.e., SketchedAMSGrad) utilizing sketching.
Our new algorithm achieves a fast convergence rate of $O(frac1sqrtnT + frac1(k/d)2 T)$ with the communication cost of $O(k log(d))$ at each iteration.
arXiv Detail & Related papers (2022-10-14T01:42:05Z) - CloudAttention: Efficient Multi-Scale Attention Scheme For 3D Point
Cloud Learning [81.85951026033787]
We set transformers in this work and incorporate them into a hierarchical framework for shape classification and part and scene segmentation.
We also compute efficient and dynamic global cross attentions by leveraging sampling and grouping at each iteration.
The proposed hierarchical model achieves state-of-the-art shape classification in mean accuracy and yields results on par with the previous segmentation methods.
arXiv Detail & Related papers (2022-07-31T21:39:15Z) - Fast Single-Core K-Nearest Neighbor Graph Computation [0.0]
This paper presents a optimized C implementation of the runtime "NN-Descent" algorithm by Wei Dong et al.
Various implementation optimizations are explained which improve performance for low-dimensional as well as high dimensional datasets.
The restriction to the l2-distance metric allows for the use of blocked distance evaluations which significantly increase performance for high dimensional datasets.
arXiv Detail & Related papers (2021-12-13T13:16:30Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - A Practical Index Structure Supporting Fr\'echet Proximity Queries Among
Trajectories [1.9335262420787858]
We present a scalable approach for range and $k$ nearest neighbor queries under computationally expensive metrics.
Based on clustering for metric indexes, we obtain a dynamic tree structure whose size is linear in the number of trajectories.
We analyze the efficiency and effectiveness of our methods with extensive experiments on diverse synthetic and real-world data sets.
arXiv Detail & Related papers (2020-05-28T04:12:43Z) - A Study of Performance of Optimal Transport [16.847501106437534]
We show that network simplex and augmenting path based algorithms can consistently outperform numerical matrix-scaling based methods.
We present a new algorithm that improves upon the classical Kuhn-Munkres algorithm.
arXiv Detail & Related papers (2020-05-03T20:37:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.