LKD-Net: Large Kernel Convolution Network for Single Image Dehazing
- URL: http://arxiv.org/abs/2209.01788v1
- Date: Mon, 5 Sep 2022 06:56:48 GMT
- Title: LKD-Net: Large Kernel Convolution Network for Single Image Dehazing
- Authors: Pinjun Luo, Guoqiang Xiao, Xinbo Gao, Song Wu
- Abstract summary: We propose a novel Large Kernel Convolution Dehaze Block (LKD Block) consisting of the Decomposition deep-wise Large Kernel Convolution Block (DLKCB) and the Channel Enhanced Feed-forward Network (CEFN)
The designed DLKCB can split the deep-wise large kernel convolution into a smaller depth-wise convolution and a depth-wise dilated convolution without introducing massive parameters and computational overhead.
Our LKD-Net dramatically outperforms the Transformer-based method Dehamer with only 1.79% #Param and 48.9% FLOPs.
- Score: 70.46392287128307
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The deep convolutional neural networks (CNNs)-based single image dehazing
methods have achieved significant success. The previous methods are devoted to
improving the network's performance by increasing the network's depth and
width. The current methods focus on increasing the convolutional kernel size to
enhance its performance by benefiting from the larger receptive field. However,
directly increasing the size of the convolutional kernel introduces a massive
amount of computational overhead and parameters. Thus, a novel Large Kernel
Convolution Dehaze Block (LKD Block) consisting of the Decomposition deep-wise
Large Kernel Convolution Block (DLKCB) and the Channel Enhanced Feed-forward
Network (CEFN) is devised in this paper. The designed DLKCB can split the
deep-wise large kernel convolution into a smaller depth-wise convolution and a
depth-wise dilated convolution without introducing massive parameters and
computational overhead. Meanwhile, the designed CEFN incorporates a channel
attention mechanism into Feed-forward Network to exploit significant channels
and enhance robustness. By combining multiple LKD Blocks and Up-Down sampling
modules, the Large Kernel Convolution Dehaze Network (LKD-Net) is conducted.
The evaluation results demonstrate the effectiveness of the designed DLKCB and
CEFN, and our LKD-Net outperforms the state-of-the-art. On the SOTS indoor
dataset, our LKD-Net dramatically outperforms the Transformer-based method
Dehamer with only 1.79% #Param and 48.9% FLOPs. The source code of our LKD-Net
is available at https://github.com/SWU-CS-MediaLab/LKD-Net.
Related papers
- SGLP: A Similarity Guided Fast Layer Partition Pruning for Compressing Large Deep Models [19.479746878680707]
Layer pruning is a potent approach to reduce network size and improve computational efficiency.
We propose a Similarity Guided fast Layer Partition pruning for compressing large deep models.
Our method outperforms the state-of-the-art methods in both accuracy and computational efficiency.
arXiv Detail & Related papers (2024-10-14T04:01:08Z) - LSK3DNet: Towards Effective and Efficient 3D Perception with Large Sparse Kernels [62.31333169413391]
Large Sparse Kernel 3D Neural Network (LSK3DNet)
Our method comprises two core components: Spatial-wise Dynamic Sparsity (SDS) and Channel-wise Weight Selection (CWS)
arXiv Detail & Related papers (2024-03-22T12:54:33Z) - HyPHEN: A Hybrid Packing Method and Optimizations for Homomorphic
Encryption-Based Neural Networks [7.642103082787977]
Convolutional neural network (CNN) inference using fully homomorphic encryption (FHE) is a promising private inference (PI) solution.
We present HyPHEN, a deep HCNN construction that incorporates novel convolution algorithms and data packing methods.
As a result, HyPHEN brings the latency of HCNN CIFAR-10 inference down to a practical level at 1.4 seconds (ResNet-20) and demonstrates HCNN ImageNet inference for the first time at 14.7 seconds (ResNet-18).
arXiv Detail & Related papers (2023-02-05T15:36:51Z) - IMDeception: Grouped Information Distilling Super-Resolution Network [7.6146285961466]
Single-Image-Super-Resolution (SISR) is a classical computer vision problem that has benefited from the recent advancements in deep learning methods.
In this work, we propose the Global Progressive Refinement Module (GPRM) as a less parameter-demanding alternative to the IIC module for feature aggregation.
We also propose Grouped Information Distilling Blocks (GIDB) to further decrease the number of parameters and floating point operations persecond (FLOPS)
Experiments reveal that the proposed network performs on par with state-of-the-art models despite having a limited number of parameters and FLOPS
arXiv Detail & Related papers (2022-04-25T06:43:45Z) - Searching for Network Width with Bilaterally Coupled Network [75.43658047510334]
We introduce a new supernet called Bilaterally Coupled Network (BCNet) to address this issue.
In BCNet, each channel is fairly trained and responsible for the same amount of network widths, thus each network width can be evaluated more accurately.
We propose the first open-source width benchmark on macro structures named Channel-Bench-Macro for the better comparison of width search algorithms.
arXiv Detail & Related papers (2022-03-25T15:32:46Z) - Rapid training of deep neural networks without skip connections or
normalization layers using Deep Kernel Shaping [46.083745557823164]
We identify the main pathologies present in deep networks that prevent them from training fast and generalizing to unseen data.
We show how these can be avoided by carefully controlling the "shape" of the network's kernel function.
arXiv Detail & Related papers (2021-10-05T00:49:36Z) - BCNet: Searching for Network Width with Bilaterally Coupled Network [56.14248440683152]
We introduce a new supernet called Bilaterally Coupled Network (BCNet) to address this issue.
In BCNet, each channel is fairly trained and responsible for the same amount of network widths, thus each network width can be evaluated more accurately.
Our method achieves state-of-the-art or competing performance over other baseline methods.
arXiv Detail & Related papers (2021-05-21T18:54:03Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z) - Finite Versus Infinite Neural Networks: an Empirical Study [69.07049353209463]
kernel methods outperform fully-connected finite-width networks.
Centered and ensembled finite networks have reduced posterior variance.
Weight decay and the use of a large learning rate break the correspondence between finite and infinite networks.
arXiv Detail & Related papers (2020-07-31T01:57:47Z) - DRU-net: An Efficient Deep Convolutional Neural Network for Medical
Image Segmentation [2.3574651879602215]
Residual network (ResNet) and densely connected network (DenseNet) have significantly improved the training efficiency and performance of deep convolutional neural networks (DCNNs)
We propose an efficient network architecture by considering advantages of both networks.
arXiv Detail & Related papers (2020-04-28T12:16:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.