Hierarchical Spherical CNNs with Lifting-based Adaptive Wavelets for
Pooling and Unpooling
- URL: http://arxiv.org/abs/2205.15571v1
- Date: Tue, 31 May 2022 07:23:42 GMT
- Title: Hierarchical Spherical CNNs with Lifting-based Adaptive Wavelets for
Pooling and Unpooling
- Authors: Mingxing Xu, Chenglin Li, Wenrui Dai, Siheng Chen, Junni Zou, Pascal
Frossard, Hongkai Xiong
- Abstract summary: We propose a novel framework of hierarchical convolutional neural networks (HS-CNNs) with a lifting structure to learn adaptive spherical wavelets for pooling and unpooling.
LiftHS-CNN ensures a more efficient hierarchical feature learning for both image- and pixel-level tasks.
- Score: 101.72318949104627
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pooling and unpooling are two essential operations in constructing
hierarchical spherical convolutional neural networks (HS-CNNs) for
comprehensive feature learning in the spherical domain. Most existing models
employ downsampling-based pooling, which will inevitably incur information loss
and cannot adapt to different spherical signals and tasks. Besides, the
preserved information after pooling cannot be well restored by the subsequent
unpooling to characterize the desirable features for a task. In this paper, we
propose a novel framework of HS-CNNs with a lifting structure to learn adaptive
spherical wavelets for pooling and unpooling, dubbed LiftHS-CNN, which ensures
a more efficient hierarchical feature learning for both image- and pixel-level
tasks. Specifically, adaptive spherical wavelets are learned with a lifting
structure that consists of trainable lifting operators (i.e., update and
predict operators). With this learnable lifting structure, we can adaptively
partition a signal into two sub-bands containing low- and high-frequency
components, respectively, and thus generate a better down-scaled representation
for pooling by preserving more information in the low-frequency sub-band. The
update and predict operators are parameterized with graph-based attention to
jointly consider the signal's characteristics and the underlying geometries. We
further show that particular properties are promised by the learned wavelets,
ensuring the spatial-frequency localization for better exploiting the signal's
correlation in both spatial and frequency domains. We then propose an unpooling
operation that is invertible to the lifting-based pooling, where an inverse
wavelet transform is performed by using the learned lifting operators to
restore an up-scaled representation. Extensive empirical evaluations on various
spherical domain tasks validate the superiority of the proposed LiftHS-CNN.
Related papers
- Wav-KAN: Wavelet Kolmogorov-Arnold Networks [3.38220960870904]
Wav-KAN is an innovative neural network architecture that leverages the Wavelet Kolmogorov-Arnold Networks (Wav-KAN) framework to enhance interpretability and performance.
Our results highlight the potential of Wav-KAN as a powerful tool for developing interpretable and high-performance neural networks.
arXiv Detail & Related papers (2024-05-21T14:36:16Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - ReLU Neural Networks with Linear Layers are Biased Towards Single- and Multi-Index Models [9.96121040675476]
This manuscript explores how properties of functions learned by neural networks of depth greater than two layers affect predictions.
Our framework considers a family of networks of varying depths that all have the same capacity but different representation costs.
arXiv Detail & Related papers (2023-05-24T22:10:12Z) - Investigating the Potential of Auxiliary-Classifier GANs for Image
Classification in Low Data Regimes [12.128005423388226]
We examine the potential for Auxiliary-Classifier GANs (AC-GANs) as a 'one-stop-shop' architecture for image classification.
AC-GANs show promise in image classification, achieving competitive performance with standard CNNs.
arXiv Detail & Related papers (2022-01-22T19:33:16Z) - AdaPool: Exponential Adaptive Pooling for Information-Retaining
Downsampling [82.08631594071656]
Pooling layers are essential building blocks of Convolutional Neural Networks (CNNs)
We propose an adaptive and exponentially weighted pooling method named adaPool.
We demonstrate how adaPool improves the preservation of detail through a range of tasks including image and video classification and object detection.
arXiv Detail & Related papers (2021-11-01T08:50:37Z) - Learnable Discrete Wavelet Pooling (LDW-Pooling) For Convolutional
Networks [33.45407848136399]
We introduce the Learning Discrete Wavelet Pooling (LDW-Pooling) that can be applied universally to replace standard pooling operations.
LDW-Pooling is effective and efficient when compared with other state-of-the-art pooling techniques such as WaveletPooling and LiftPooling.
arXiv Detail & Related papers (2021-09-13T08:02:38Z) - Compressing Deep ODE-Nets using Basis Function Expansions [105.05435207079759]
We consider formulations of the weights as continuous-depth functions using linear combinations of basis functions.
This perspective allows us to compress the weights through a change of basis, without retraining, while maintaining near state-of-the-art performance.
In turn, both inference time and the memory footprint are reduced, enabling quick and rigorous adaptation between computational environments.
arXiv Detail & Related papers (2021-06-21T03:04:51Z) - Data-Driven Learning of Geometric Scattering Networks [74.3283600072357]
We propose a new graph neural network (GNN) module based on relaxations of recently proposed geometric scattering transforms.
Our learnable geometric scattering (LEGS) module enables adaptive tuning of the wavelets to encourage band-pass features to emerge in learned representations.
arXiv Detail & Related papers (2020-10-06T01:20:27Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - Multi Layer Neural Networks as Replacement for Pooling Operations [13.481518628796692]
We show that one perceptron can already be used effectively as a pooling operation without increasing the complexity of the model.
We compare our approach to tensor convolution with strides as a pooling operation and show that our approach is both effective and reduces complexity.
arXiv Detail & Related papers (2020-06-12T07:08:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.