Hierarchical Partitioning Forecaster
- URL: http://arxiv.org/abs/2305.13063v1
- Date: Mon, 22 May 2023 14:25:46 GMT
- Title: Hierarchical Partitioning Forecaster
- Authors: Christopher Mattern
- Abstract summary: We consider a new family of algorithms for sequential prediction, Hierarchical Partitioning Forecasters (HPFs)
Our goal is to provide appealing theoretical - regret guarantees on a powerful model class.
- Score: 0.571097144710995
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we consider a new family of algorithms for sequential
prediction, Hierarchical Partitioning Forecasters (HPFs). Our goal is to
provide appealing theoretical - regret guarantees on a powerful model class -
and practical - empirical performance comparable to deep networks - properties
at the same time. We built upon three principles: hierarchically partitioning
the feature space into sub-spaces, blending forecasters specialized to each
sub-space and learning HPFs via local online learning applied to these
individual forecasters. Following these principles allows us to obtain regret
guarantees, where Constant Partitioning Forecasters (CPFs) serve as competitor.
A CPF partitions the feature space into sub-spaces and predicts with a fixed
forecaster per sub-space. Fixing a hierarchical partition $\mathcal H$ and
considering any CPF with a partition that can be constructed using elements of
$\mathcal H$ we provide two guarantees: first, a generic one that unveils how
local online learning determines regret of learning the entire HPF online;
second, a concrete instance that considers HPF with linear forecasters (LHPF)
and exp-concave losses where we obtain $O(k \log T)$ regret for sequences of
length $T$ where $k$ is a measure of complexity for the competing CPF. Finally,
we provide experiments that compare LHPF to various baselines, including state
of the art deep learning models, in precipitation nowcasting. Our results
indicate that LHPF is competitive in various settings.
Related papers
- Smoothed Normalization for Efficient Distributed Private Optimization [54.197255548244705]
Federated learning enables machine learning models with privacy of participants.
There is no differentially private distributed method for training, non-feedback problems.
We introduce a new distributed algorithm $alpha$-$sf NormEC$ with provable convergence guarantees.
arXiv Detail & Related papers (2025-02-19T07:10:32Z) - Smoothed Analysis for Learning Concepts with Low Intrinsic Dimension [17.485243410774814]
In traditional models of supervised learning, the goal of a learner is to output a hypothesis that is competitive (to within $epsilon$) of the best fitting concept from some class.
We introduce a smoothed-analysis framework that requires a learner to compete only with the best agnostic.
We obtain the first algorithm forally learning intersections of $k$-halfspaces in time.
arXiv Detail & Related papers (2024-07-01T04:58:36Z) - Rethinking Few-shot 3D Point Cloud Semantic Segmentation [62.80639841429669]
This paper revisits few-shot 3D point cloud semantic segmentation (FS-PCS)
We focus on two significant issues in the state-of-the-art: foreground leakage and sparse point distribution.
To address these issues, we introduce a standardized FS-PCS setting, upon which a new benchmark is built.
arXiv Detail & Related papers (2024-03-01T15:14:47Z) - Inferring Dynamic Networks from Marginals with Iterative Proportional Fitting [57.487936697747024]
A common network inference problem, arising from real-world data constraints, is how to infer a dynamic network from its time-aggregated adjacency matrix.
We introduce a principled algorithm that guarantees IPF converges under minimal changes to the network structure.
arXiv Detail & Related papers (2024-02-28T20:24:56Z) - ParsNets: A Parsimonious Orthogonal and Low-Rank Linear Networks for
Zero-Shot Learning [22.823915322924304]
This paper provides a novel parsimonious yet efficient design for zero-shot learning (ZSL), dubbed ParsNets, to achieve equivalent or even better performance against existing deep models.
To facilitate the generalization of local linearities, we construct a maximal margin geometry on the learned features by enforcing low-rank constraints on intra-class samples and high-rank constraints on inter-class samples.
To enhance the model's adaptability and counterbalance over/under-fittings in ZSL, a set of sample-wise indicators is employed to select a sparse subset from these base linear networks to form a composite
arXiv Detail & Related papers (2023-12-15T11:32:11Z) - Federated Learning over Hierarchical Wireless Networks: Training Latency Minimization via Submodel Partitioning [15.311309249848739]
Hierarchical independent submodel training (HIST) is a new FL methodology that aims to address these issues in hierarchical cloud-edge-client networks.
We demonstrate how HIST can be augmented with over-the-air computation (AirComp) to further enhance the efficiency of the model aggregation over the edge cells.
arXiv Detail & Related papers (2023-10-27T04:42:59Z) - GIFD: A Generative Gradient Inversion Method with Feature Domain
Optimization [52.55628139825667]
Federated Learning (FL) has emerged as a promising distributed machine learning framework to preserve clients' privacy.
Recent studies find that an attacker can invert the shared gradients and recover sensitive data against an FL system by leveraging pre-trained generative adversarial networks (GAN) as prior knowledge.
We propose textbfGradient textbfInversion over textbfFeature textbfDomains (GIFD), which disassembles the GAN model and searches the feature domains of the intermediate layers.
arXiv Detail & Related papers (2023-08-09T04:34:21Z) - Subspace Learning Machine (SLM): Methodology and Performance [28.98486923400986]
Subspace learning machine (SLM) is a new classification model inspired by feedforward multilayer perceptron (FF-MLP), decision tree (DT) and extreme learning machine (ELM)
SLM first identifies a discriminant subspace, $S0$, by examining the discriminant power of each input feature.
It uses probabilistic projections of features in $S0$ to yield 1D subspaces and finds the optimal partition for each of them.
arXiv Detail & Related papers (2022-05-11T06:44:51Z) - Partition of unity networks: deep hp-approximation [0.0]
We propose partition of unity networks (POUnets) which incorporate these elements directly into the architecture.
POUnets yield hp-convergence for smooth functions and consistently outperform piecewise functions with large numbers of discontinuities.
arXiv Detail & Related papers (2021-01-27T08:26:11Z) - Joint and Progressive Subspace Analysis (JPSA) with Spatial-Spectral
Manifold Alignment for Semi-Supervised Hyperspectral Dimensionality Reduction [48.73525876467408]
We propose a novel technique for hyperspectral subspace analysis.
The technique is called joint and progressive subspace analysis (JPSA)
Experiments are conducted to demonstrate the superiority and effectiveness of the proposed JPSA on two widely-used hyperspectral datasets.
arXiv Detail & Related papers (2020-09-21T16:29:59Z) - Learning Halfspaces with Tsybakov Noise [50.659479930171585]
We study the learnability of halfspaces in the presence of Tsybakov noise.
We give an algorithm that achieves misclassification error $epsilon$ with respect to the true halfspace.
arXiv Detail & Related papers (2020-06-11T14:25:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.