Learning from Small Samples: Transformation-Invariant SVMs with
Composition and Locality at Multiple Scales
- URL: http://arxiv.org/abs/2109.12784v2
- Date: Tue, 28 Sep 2021 05:31:13 GMT
- Title: Learning from Small Samples: Transformation-Invariant SVMs with
Composition and Locality at Multiple Scales
- Authors: Tao Liu, P. R. Kumar, Xi Liu
- Abstract summary: This paper shows how to incorporate into support-vector machines (SVMs) those properties that have made convolutional neural networks (CNNs) successful.
- Score: 11.210266084524998
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motivated by the problem of learning when the number of training samples is
small, this paper shows how to incorporate into support-vector machines (SVMs)
those properties that have made convolutional neural networks (CNNs)
successful. Particularly important is the ability to incorporate domain
knowledge of invariances, e.g., translational invariance of images. Kernels
based on the \textit{minimum} distance over a group of transformations, which
corresponds to defining similarity as the \textit{best} over the possible
transformations, are not generally positive definite. Perhaps it is for this
reason that they have neither previously been experimentally tested for their
performance nor studied theoretically. Instead, previous attempts have employed
kernels based on the \textit{average} distance over a group of transformations,
which are trivially positive definite, but which generally yield both poor
margins as well as poor performance, as we show. We address this lacuna and
show that positive definiteness indeed holds \textit{with high probability} for
kernels based on the minimum distance in the small training sample set regime
of interest, and that they do yield the best results in that regime. Another
important property of CNNs is their ability to incorporate local features at
multiple spatial scales, e.g., through max pooling. A third important property
is their ability to provide the benefits of composition through the
architecture of multiple layers. We show how these additional properties can
also be embedded into SVMs. We verify through experiments on widely available
image sets that the resulting SVMs do provide superior accuracy in comparison
to well-established deep neural network (DNN) benchmarks for small sample
sizes.
Related papers
- SO(2) and O(2) Equivariance in Image Recognition with
Bessel-Convolutional Neural Networks [63.24965775030674]
This work presents the development of Bessel-convolutional neural networks (B-CNNs)
B-CNNs exploit a particular decomposition based on Bessel functions to modify the key operation between images and filters.
Study is carried out to assess the performances of B-CNNs compared to other methods.
arXiv Detail & Related papers (2023-04-18T18:06:35Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - Focal Sparse Convolutional Networks for 3D Object Detection [121.45950754511021]
We introduce two new modules to enhance the capability of Sparse CNNs.
They are focal sparse convolution (Focals Conv) and its multi-modal variant of focal sparse convolution with fusion.
For the first time, we show that spatially learnable sparsity in sparse convolution is essential for sophisticated 3D object detection.
arXiv Detail & Related papers (2022-04-26T17:34:10Z) - Improving the Sample-Complexity of Deep Classification Networks with
Invariant Integration [77.99182201815763]
Leveraging prior knowledge on intraclass variance due to transformations is a powerful method to improve the sample complexity of deep neural networks.
We propose a novel monomial selection algorithm based on pruning methods to allow an application to more complex problems.
We demonstrate the improved sample complexity on the Rotated-MNIST, SVHN and CIFAR-10 datasets.
arXiv Detail & Related papers (2022-02-08T16:16:11Z) - Exploiting Invariance in Training Deep Neural Networks [4.169130102668252]
Inspired by two basic mechanisms in animal visual systems, we introduce a feature transform technique that imposes invariance properties in the training of deep neural networks.
The resulting algorithm requires less parameter tuning, trains well with an initial learning rate 1.0, and easily generalizes to different tasks.
Tested on ImageNet, MS COCO, and Cityscapes datasets, our proposed technique requires fewer iterations to train, surpasses all baselines by a large margin, seamlessly works on both small and large batch size training, and applies to different computer vision tasks of image classification, object detection, and semantic segmentation.
arXiv Detail & Related papers (2021-03-30T19:18:31Z) - Learning Invariances in Neural Networks [51.20867785006147]
We show how to parameterize a distribution over augmentations and optimize the training loss simultaneously with respect to the network parameters and augmentation parameters.
We can recover the correct set and extent of invariances on image classification, regression, segmentation, and molecular property prediction from a large space of augmentations.
arXiv Detail & Related papers (2020-10-22T17:18:48Z) - Margin-Based Regularization and Selective Sampling in Deep Neural
Networks [7.219077740523683]
We derive a new margin-based regularization formulation, termed multi-margin regularization (MMR) for deep neural networks (DNNs)
We show improved empirical results on CIFAR10, CIFAR100 and ImageNet using state-of-the-art convolutional neural networks (CNNs) and BERT-BASE architecture for the MNLI, QQP, QNLI, MRPC, SST-2 and RTE benchmarks.
arXiv Detail & Related papers (2020-09-13T15:06:42Z) - Embedding Propagation: Smoother Manifold for Few-Shot Classification [131.81692677836202]
We propose to use embedding propagation as an unsupervised non-parametric regularizer for manifold smoothing in few-shot classification.
We empirically show that embedding propagation yields a smoother embedding manifold.
We show that embedding propagation consistently improves the accuracy of the models in multiple semi-supervised learning scenarios by up to 16% points.
arXiv Detail & Related papers (2020-03-09T13:51:09Z) - On the Texture Bias for Few-Shot CNN Segmentation [21.349705243254423]
Convolutional Neural Networks (CNNs) are driven by shapes to perform visual recognition tasks.
Recent evidence suggests texture bias in CNNs provides higher performing models when learning on large labeled training datasets.
We propose a novel architecture that integrates a set of Difference of Gaussians (DoG) to attenuate high-frequency local components in the feature space.
arXiv Detail & Related papers (2020-03-09T11:55:47Z) - Convolutional Support Vector Machine [1.5990720051907859]
This paper proposes a novel convolutional SVM (CSVM) that has both advantages of CNN and SVM to improve the accuracy and effectiveness of mining smaller datasets.
To evaluate the performance of the proposed CSVM, experiments were conducted to test five well-known benchmark databases for the classification problem.
arXiv Detail & Related papers (2020-02-11T11:23:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.