Comparison of Methods Generalizing Max- and Average-Pooling
- URL: http://arxiv.org/abs/2103.01746v1
- Date: Tue, 2 Mar 2021 14:26:51 GMT
- Title: Comparison of Methods Generalizing Max- and Average-Pooling
- Authors: Florentin Bieder, Robin Sandk\"uhler, Philippe C. Cattin
- Abstract summary: Max- and average-pooling are the most popular methods for downsampling in convolutional neural networks.
In this paper, we compare different pooling methods that generalize both max- and average-pooling.
The results show that none of the more sophisticated methods perform significantly better in this classification task than standard max- or average-pooling.
- Score: 1.693200946453174
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Max- and average-pooling are the most popular pooling methods for
downsampling in convolutional neural networks. In this paper, we compare
different pooling methods that generalize both max- and average-pooling.
Furthermore, we propose another method based on a smooth approximation of the
maximum function and put it into context with related methods. For the
comparison, we use a VGG16 image classification network and train it on a large
dataset of natural high-resolution images (Google Open Images v5). The results
show that none of the more sophisticated methods perform significantly better
in this classification task than standard max- or average-pooling.
Related papers
- A Gauss-Newton Approach for Min-Max Optimization in Generative Adversarial Networks [7.217857709620766]
A novel first-order method is proposed for training generative adversarial networks (GANs)
It modifies the Gauss-Newton method to approximate the min-max Hessian and uses the Sherman-Morrison inversion formula to calculate the inverse.
Our method is capable of generating high-fidelity images with greater diversity across multiple datasets.
arXiv Detail & Related papers (2024-04-10T17:08:46Z) - MB-RACS: Measurement-Bounds-based Rate-Adaptive Image Compressed Sensing Network [65.1004435124796]
We propose a Measurement-Bounds-based Rate-Adaptive Image Compressed Sensing Network (MB-RACS) framework.
Our experiments demonstrate that the proposed MB-RACS method surpasses current leading methods.
arXiv Detail & Related papers (2024-01-19T04:40:20Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - The Theoretical Expressiveness of Maxpooling [4.028503203417233]
We develop a theoretical framework analyzing ReLU based approximations to max pooling.
We find that max pooling cannot be efficiently replicated using ReLU activations.
We conclude that the main cause of a difference between max pooling and an optimal approximation, can be overcome with other architectural decisions.
arXiv Detail & Related papers (2022-03-02T10:45:53Z) - AdaPool: Exponential Adaptive Pooling for Information-Retaining
Downsampling [82.08631594071656]
Pooling layers are essential building blocks of Convolutional Neural Networks (CNNs)
We propose an adaptive and exponentially weighted pooling method named adaPool.
We demonstrate how adaPool improves the preservation of detail through a range of tasks including image and video classification and object detection.
arXiv Detail & Related papers (2021-11-01T08:50:37Z) - PixelPyramids: Exact Inference Models from Lossless Image Pyramids [58.949070311990916]
Pixel-Pyramids is a block-autoregressive approach with scale-specific representations to encode the joint distribution of image pixels.
It yields state-of-the-art results for density estimation on various image datasets, especially for high-resolution data.
For CelebA-HQ 1024 x 1024, we observe that the density estimates are improved to 44% of the baseline despite sampling speeds superior even to easily parallelizable flow-based models.
arXiv Detail & Related papers (2021-10-17T10:47:29Z) - Maximal function pooling with applications [4.446564162927513]
Maxfun pooling is inspired by the Hardy-Littlewood maximal function.
It is presented as a viable alternative to some of the most popular pooling functions, such as max pooling and average pooling.
We demonstrate the features of maxfun pooling with two applications: first in the context of convolutional sparse coding, and then for image classification.
arXiv Detail & Related papers (2021-03-01T20:30:04Z) - Unsupervised Learning of Visual Features by Contrasting Cluster
Assignments [57.33699905852397]
We propose an online algorithm, SwAV, that takes advantage of contrastive methods without requiring to compute pairwise comparisons.
Our method simultaneously clusters the data while enforcing consistency between cluster assignments.
Our method can be trained with large and small batches and can scale to unlimited amounts of data.
arXiv Detail & Related papers (2020-06-17T14:00:42Z) - Clustering Binary Data by Application of Combinatorial Optimization
Heuristics [52.77024349608834]
We study clustering methods for binary data, first defining aggregation criteria that measure the compactness of clusters.
Five new and original methods are introduced, using neighborhoods and population behavior optimization metaheuristics.
From a set of 16 data tables generated by a quasi-Monte Carlo experiment, a comparison is performed for one of the aggregations using L1 dissimilarity, with hierarchical clustering, and a version of k-means: partitioning around medoids or PAM.
arXiv Detail & Related papers (2020-01-06T23:33:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.