Pseudo-Inverted Bottleneck Convolution for DARTS Search Space
- URL: http://arxiv.org/abs/2301.01286v3
- Date: Sun, 19 Mar 2023 00:49:26 GMT
- Title: Pseudo-Inverted Bottleneck Convolution for DARTS Search Space
- Authors: Arash Ahmadian, Louis S.P. Liu, Yue Fei, Konstantinos N. Plataniotis,
Mahdi S. Hosseini
- Abstract summary: We introduce the Pseudo-Inverted Bottleneck (PIBConv) block intending to reduce the computational footprint of the inverted bottleneck block proposed in ConvNeXt.
Our proposed architecture is much less sensitive to evaluation layer count and outperforms a DARTS network with similar size significantly, at layer counts as small as 2.
- Score: 35.50068534514941
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Differentiable Architecture Search (DARTS) has attracted considerable
attention as a gradient-based neural architecture search method. Since the
introduction of DARTS, there has been little work done on adapting the action
space based on state-of-art architecture design principles for CNNs. In this
work, we aim to address this gap by incrementally augmenting the DARTS search
space with micro-design changes inspired by ConvNeXt and studying the trade-off
between accuracy, evaluation layer count, and computational cost. We introduce
the Pseudo-Inverted Bottleneck Conv (PIBConv) block intending to reduce the
computational footprint of the inverted bottleneck block proposed in ConvNeXt.
Our proposed architecture is much less sensitive to evaluation layer count and
outperforms a DARTS network with similar size significantly, at layer counts as
small as 2. Furthermore, with less layers, not only does it achieve higher
accuracy with lower computational footprint (measured in GMACs) and parameter
count, GradCAM comparisons show that our network can better detect distinctive
features of target objects compared to DARTS. Code is available from
https://github.com/mahdihosseini/PIBConv.
Related papers
- DASS: Differentiable Architecture Search for Sparse neural networks [0.5735035463793009]
We find that the architectures designed for dense networks by differentiable architecture search methods are ineffective when pruning mechanisms are applied to them.
In this paper, we propose a new method to search for sparsity-friendly neural architectures.
We do this by adding two new sparse operations to the search space and modifying the search objective.
arXiv Detail & Related papers (2022-07-14T14:53:50Z) - EResFD: Rediscovery of the Effectiveness of Standard Convolution for
Lightweight Face Detection [13.357235715178584]
We re-examine the effectiveness of the standard convolutional block as a lightweight backbone architecture for face detection.
We show that heavily channel-pruned standard convolution layers can achieve better accuracy and inference speed.
Our proposed detector EResFD obtained 80.4% mAP on WIDER FACE Hard subset which only takes 37.7 ms for VGA image inference on CPU.
arXiv Detail & Related papers (2022-04-04T02:30:43Z) - MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning [72.80896338009579]
We find that the memory bottleneck is due to the imbalanced memory distribution in convolutional neural network (CNN) designs.
We propose a generic patch-by-patch inference scheduling, which significantly cuts down the peak memory.
We automate the process with neural architecture search to jointly optimize the neural architecture and inference scheduling, leading to MCUNetV2.
arXiv Detail & Related papers (2021-10-28T17:58:45Z) - D-DARTS: Distributed Differentiable Architecture Search [75.12821786565318]
Differentiable ARchiTecture Search (DARTS) is one of the most trending Neural Architecture Search (NAS) methods.
We propose D-DARTS, a novel solution that addresses this problem by nesting several neural networks at cell-level.
arXiv Detail & Related papers (2021-08-20T09:07:01Z) - iDARTS: Differentiable Architecture Search with Stochastic Implicit
Gradients [75.41173109807735]
Differentiable ARchiTecture Search (DARTS) has recently become the mainstream of neural architecture search (NAS)
We tackle the hypergradient computation in DARTS based on the implicit function theorem.
We show that the architecture optimisation with the proposed method, named iDARTS, is expected to converge to a stationary point.
arXiv Detail & Related papers (2021-06-21T00:44:11Z) - Rethinking Bottleneck Structure for Efficient Mobile Network Design [154.47657111869552]
The inverted residual block is dominating architecture design for mobile networks recently.
We propose to flip the structure and present a novel bottleneck design, called the sandglass block, that performs identity mapping and spatial transformation at higher dimensions.
In ImageNet classification, by simply replacing the inverted residual block with our sandglass block without increasing parameters and computation, the classification accuracy can be improved by more than 1.7% over MobileNetV2.
arXiv Detail & Related papers (2020-07-05T08:55:26Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.