Few-shot Learning as Cluster-induced Voronoi Diagrams: A Geometric
Approach
- URL: http://arxiv.org/abs/2202.02471v1
- Date: Sat, 5 Feb 2022 02:52:06 GMT
- Title: Few-shot Learning as Cluster-induced Voronoi Diagrams: A Geometric
Approach
- Authors: Chunwei Ma, Ziyun Huang, Mingchen Gao and Jinhui Xu
- Abstract summary: Cluster-induced Voronoi Diagram (CIVD) improves the accuracy and robustness of few-shot learning.
Our CIVD-based workflow enables us to achieve new state-of-the-art results on mini-ImageNet, CUB, and tiered-ImagenNet datasets.
- Score: 12.382578792491747
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning (FSL) is the process of rapid generalization from abundant
base samples to inadequate novel samples. Despite extensive research in recent
years, FSL is still not yet able to generate satisfactory solutions for a wide
range of real-world applications. To confront this challenge, we study the FSL
problem from a geometric point of view in this paper. One observation is that
the widely embraced ProtoNet model is essentially a Voronoi Diagram (VD) in the
feature space. We retrofit it by making use of a recent advance in
computational geometry called Cluster-induced Voronoi Diagram (CIVD). Starting
from the simplest nearest neighbor model, CIVD gradually incorporates
cluster-to-point and then cluster-to-cluster relationships for space
subdivision, which is used to improve the accuracy and robustness at multiple
stages of FSL. Specifically, we use CIVD (1) to integrate parametric and
nonparametric few-shot classifiers; (2) to combine feature representation and
surrogate representation; (3) and to leverage feature-level,
transformation-level, and geometry-level heterogeneities for a better ensemble.
Our CIVD-based workflow enables us to achieve new state-of-the-art results on
mini-ImageNet, CUB, and tiered-ImagenNet datasets, with ${\sim}2\%{-}5\%$
improvements upon the next best. To summarize, CIVD provides a mathematically
elegant and geometrically interpretable framework that compensates for extreme
data insufficiency, prevents overfitting, and allows for fast geometric
ensemble for thousands of individual VD. These together make FSL stronger.
Related papers
- TTVD: Towards a Geometric Framework for Test-Time Adaptation Based on Voronoi Diagram [14.238620530634392]
Test-time adaptation (TTA) is an emerging scheme used at inference time to address this issue.
We study the TTA problem from a geometric point of view.
We propose the Test-Time adjustment by Voronoi Diagram guidance (TTVD), a novel framework that leverages the benefits of this geometric property.
arXiv Detail & Related papers (2024-12-10T23:40:07Z) - Point Cloud Denoising With Fine-Granularity Dynamic Graph Convolutional Networks [58.050130177241186]
Noise perturbations often corrupt 3-D point clouds, hindering downstream tasks such as surface reconstruction, rendering, and further processing.
This paper introduces finegranularity dynamic graph convolutional networks called GDGCN, a novel approach to denoising in 3-D point clouds.
arXiv Detail & Related papers (2024-11-21T14:19:32Z) - Language Models as Zero-shot Lossless Gradient Compressors: Towards General Neural Parameter Prior Models [56.00251589760559]
Large language models (LLMs) can act as gradient priors in a zero-shot setting.
We introduce LM-GC, a novel method that integrates LLMs with arithmetic coding.
Experiments indicate that LM-GC surpasses existing state-of-the-art lossless compression methods.
arXiv Detail & Related papers (2024-09-26T13:38:33Z) - Double-Shot 3D Shape Measurement with a Dual-Branch Network for Structured Light Projection Profilometry [14.749887303860717]
We propose a dual-branch Convolutional Neural Network (CNN)-Transformer network (PDCNet) to process different structured light (SL) modalities.
Within PDCNet, a Transformer branch is used to capture global perception in the fringe images, while a CNN branch is designed to collect local details in the speckle images.
Our method can reduce fringe order ambiguity while producing high-accuracy results on self-made datasets.
arXiv Detail & Related papers (2024-07-19T10:49:26Z) - Coarse-Fine Spectral-Aware Deformable Convolution For Hyperspectral Image Reconstruction [15.537910100051866]
We study the inverse problem of Coded Aperture Snapshot Spectral Imaging (CASSI)
We propose Coarse-Fine Spectral-Aware Deformable Convolution Network (CFSDCN)
Our CFSDCN significantly outperforms previous state-of-the-art (SOTA) methods on both simulated and real HSI datasets.
arXiv Detail & Related papers (2024-06-18T15:15:12Z) - LD-GAN: Low-Dimensional Generative Adversarial Network for Spectral
Image Generation with Variance Regularization [72.4394510913927]
Deep learning methods are state-of-the-art for spectral image (SI) computational tasks.
GANs enable diverse augmentation by learning and sampling from the data distribution.
GAN-based SI generation is challenging since the high-dimensionality nature of this kind of data hinders the convergence of the GAN training yielding to suboptimal generation.
We propose a statistical regularization to control the low-dimensional representation variance for the autoencoder training and to achieve high diversity of samples generated with the GAN.
arXiv Detail & Related papers (2023-04-29T00:25:02Z) - Faster Adaptive Federated Learning [84.38913517122619]
Federated learning has attracted increasing attention with the emergence of distributed data.
In this paper, we propose an efficient adaptive algorithm (i.e., FAFED) based on momentum-based variance reduced technique in cross-silo FL.
arXiv Detail & Related papers (2022-12-02T05:07:50Z) - Learning A 3D-CNN and Transformer Prior for Hyperspectral Image
Super-Resolution [80.93870349019332]
We propose a novel HSISR method that uses Transformer instead of CNN to learn the prior of HSIs.
Specifically, we first use the gradient algorithm to solve the HSISR model, and then use an unfolding network to simulate the iterative solution processes.
arXiv Detail & Related papers (2021-11-27T15:38:57Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.