What Deep CNNs Benefit from Global Covariance Pooling: An Optimization
Perspective
- URL: http://arxiv.org/abs/2003.11241v1
- Date: Wed, 25 Mar 2020 07:00:45 GMT
- Title: What Deep CNNs Benefit from Global Covariance Pooling: An Optimization
Perspective
- Authors: Qilong Wang, Li Zhang, Banggu Wu, Dongwei Ren, Peihua Li, Wangmeng
Zuo, Qinghua Hu
- Abstract summary: We make an attempt to understand what deep CNNs benefit from GCP in a viewpoint of optimization.
We show that GCP can make the optimization landscape more smooth and the gradients more predictive.
We conduct extensive experiments using various deep CNN models on diversified tasks, and the results provide strong support to our findings.
- Score: 102.37204254403038
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent works have demonstrated that global covariance pooling (GCP) has the
ability to improve performance of deep convolutional neural networks (CNNs) on
visual classification task. Despite considerable advance, the reasons on
effectiveness of GCP on deep CNNs have not been well studied. In this paper, we
make an attempt to understand what deep CNNs benefit from GCP in a viewpoint of
optimization. Specifically, we explore the effect of GCP on deep CNNs in terms
of the Lipschitzness of optimization loss and the predictiveness of gradients,
and show that GCP can make the optimization landscape more smooth and the
gradients more predictive. Furthermore, we discuss the connection between GCP
and second-order optimization for deep CNNs. More importantly, above findings
can account for several merits of covariance pooling for training deep CNNs
that have not been recognized previously or fully explored, including
significant acceleration of network convergence (i.e., the networks trained
with GCP can support rapid decay of learning rates, achieving favorable
performance while significantly reducing number of training epochs), stronger
robustness to distorted examples generated by image corruptions and
perturbations, and good generalization ability to different vision tasks, e.g.,
object detection and instance segmentation. We conduct extensive experiments
using various deep CNN models on diversified tasks, and the results provide
strong support to our findings.
Related papers
- Unveiling the optimization process of Physics Informed Neural Networks: How accurate and competitive can PINNs be? [0.0]
This study investigates the potential accuracy of physics-informed neural networks, contrasting their approach with previous similar works and traditional numerical methods.
We find that selecting improved optimization algorithms significantly enhances the accuracy of the results.
Simple modifications to the loss function may also improve precision, offering an additional avenue for enhancement.
arXiv Detail & Related papers (2024-05-07T11:50:25Z) - Optimizing Neural Network Scale for ECG Classification [1.8953148404648703]
We study scaling convolutional neural networks (CNNs) specifically targeting Residual neural networks (ResNet) for analyzing electrocardiograms (ECGs)
We explored and demonstrated an efficient approach to scale ResNet by examining the effects of crucial parameters, including layer depth, the number of channels, and the convolution kernel size.
Our findings provide insight into obtaining more efficient and accurate models with fewer computing resources or less time.
arXiv Detail & Related papers (2023-08-24T01:26:31Z) - Understanding and Improving Deep Graph Neural Networks: A Probabilistic
Graphical Model Perspective [22.82625446308785]
We propose a novel view for understanding graph neural networks (GNNs)
In this work, we focus on deep GNNs and propose a novel view for understanding them.
We design a more powerful GNN: coupling graph neural network (CoGNet)
arXiv Detail & Related papers (2023-01-25T12:02:12Z) - Online Adaptation of Monocular Depth Prediction with Visual SLAM [8.478040209440868]
The ability of accurate depth prediction by a CNN is a major challenge for its wide use in practical visual SLAM applications.
We propose a novel online adaptation framework consisting of two complementary processes to fine-tune the depth prediction.
Experimental results on both benchmark datasets and a real robot in our own experimental environments show that our proposed method improves the SLAM reconstruction accuracy.
arXiv Detail & Related papers (2021-11-07T14:20:35Z) - CAP: Co-Adversarial Perturbation on Weights and Features for Improving
Generalization of Graph Neural Networks [59.692017490560275]
Adversarial training has been widely demonstrated to improve model's robustness against adversarial attacks.
It remains unclear how the adversarial training could improve the generalization abilities of GNNs in the graph analytics problem.
We construct the co-adversarial perturbation (CAP) optimization problem in terms of weights and features, and design the alternating adversarial perturbation algorithm to flatten the weight and feature loss landscapes alternately.
arXiv Detail & Related papers (2021-10-28T02:28:13Z) - Fusion of CNNs and statistical indicators to improve image
classification [65.51757376525798]
Convolutional Networks have dominated the field of computer vision for the last ten years.
Main strategy to prolong this trend relies on further upscaling networks in size.
We hypothesise that adding heterogeneous sources of information may be more cost-effective to a CNN than building a bigger network.
arXiv Detail & Related papers (2020-12-20T23:24:31Z) - Dynamic Hierarchical Mimicking Towards Consistent Optimization
Objectives [73.15276998621582]
We propose a generic feature learning mechanism to advance CNN training with enhanced generalization ability.
Partially inspired by DSN, we fork delicately designed side branches from the intermediate layers of a given neural network.
Experiments on both category and instance recognition tasks demonstrate the substantial improvements of our proposed method.
arXiv Detail & Related papers (2020-03-24T09:56:13Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.