Searching a High-Performance Feature Extractor for Text Recognition
Network
- URL: http://arxiv.org/abs/2209.13139v1
- Date: Tue, 27 Sep 2022 03:49:04 GMT
- Title: Searching a High-Performance Feature Extractor for Text Recognition
Network
- Authors: Hui Zhang, Quanming Yao, James T. Kwok, Xiang Bai
- Abstract summary: We design a domain-specific search space by exploring principles for having good feature extractors.
As the space is huge and complexly structured, no existing NAS algorithms can be applied.
We propose a two-stage algorithm to effectively search in the space.
- Score: 92.12492627169108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Feature extractor plays a critical role in text recognition (TR), but
customizing its architecture is relatively less explored due to expensive
manual tweaking. In this work, inspired by the success of neural architecture
search (NAS), we propose to search for suitable feature extractors. We design a
domain-specific search space by exploring principles for having good feature
extractors. The space includes a 3D-structured space for the spatial model and
a transformed-based space for the sequential model. As the space is huge and
complexly structured, no existing NAS algorithms can be applied. We propose a
two-stage algorithm to effectively search in the space. In the first stage, we
cut the space into several blocks and progressively train each block with the
help of an auxiliary head. We introduce the latency constraint into the second
stage and search sub-network from the trained supernet via natural gradient
descent. In experiments, a series of ablation studies are performed to better
understand the designed space, search algorithm, and searched architectures. We
also compare the proposed method with various state-of-the-art ones on both
hand-written and scene TR tasks. Extensive results show that our approach can
achieve better recognition performance with less latency.
Related papers
- LISSNAS: Locality-based Iterative Search Space Shrinkage for Neural
Architecture Search [30.079267927860347]
We propose an automated algorithm that shrinks a large space into a diverse, small search space with SOTA search performance.
Our method achieves a SOTA Top-1 accuracy of 77.6% in ImageNet under mobile constraints, best-in-class Kendal-Tau, architectural diversity, and search space size.
arXiv Detail & Related papers (2023-07-06T16:28:51Z) - AutoSpace: Neural Architecture Search with Less Human Interference [84.42680793945007]
Current neural architecture search (NAS) algorithms still require expert knowledge and effort to design a search space for network construction.
We propose a novel differentiable evolutionary framework named AutoSpace, which evolves the search space to an optimal one.
With the learned search space, the performance of recent NAS algorithms can be improved significantly compared with using previously manually designed spaces.
arXiv Detail & Related papers (2021-03-22T13:28:56Z) - Evolving Search Space for Neural Architecture Search [70.71153433676024]
We present a Neural Search-space Evolution (NSE) scheme that amplifies the results from the previous effort by maintaining an optimized search space subset.
We achieve 77.3% top-1 retrain accuracy on ImageNet with 333M FLOPs, which yielded a state-of-the-art performance.
When the latency constraint is adopted, our result also performs better than the previous best-performing mobile models with a 77.9% Top-1 retrain accuracy.
arXiv Detail & Related papers (2020-11-22T01:11:19Z) - Representation Sharing for Fast Object Detector Search and Beyond [38.18583590914755]
We propose Fast And Diverse (FAD) to better explore the optimal configuration of receptive fields and convolution types in the sub-networks for one-stage detectors.
FAD achieves prominent improvements on two types of one-stage detectors with various backbones.
arXiv Detail & Related papers (2020-07-23T15:39:44Z) - GOLD-NAS: Gradual, One-Level, Differentiable [100.12492801459105]
We propose a novel algorithm named Gradual One-Level Differentiable Neural Architecture Search (GOLD-NAS)
It introduces a variable resource constraint to one-level optimization so that the weak operators are gradually pruned out from the super-network.
arXiv Detail & Related papers (2020-07-07T10:37:49Z) - DrNAS: Dirichlet Neural Architecture Search [88.56953713817545]
We treat the continuously relaxed architecture mixing weight as random variables, modeled by Dirichlet distribution.
With recently developed pathwise derivatives, the Dirichlet parameters can be easily optimized with gradient-based generalization.
To alleviate the large memory consumption of differentiable NAS, we propose a simple yet effective progressive learning scheme.
arXiv Detail & Related papers (2020-06-18T08:23:02Z) - Progressive Automatic Design of Search Space for One-Shot Neural
Architecture Search [15.017964136568061]
It has been observed that a model with higher one-shot model accuracy does not necessarily perform better when stand-alone trained.
We propose Progressive Automatic Design of search space, named PAD-NAS.
In this way, PAD-NAS can automatically design the operations for each layer and achieve a trade-off between search space quality and model diversity.
arXiv Detail & Related papers (2020-05-15T14:21:07Z) - AutoSTR: Efficient Backbone Search for Scene Text Recognition [80.7290173000068]
Scene text recognition (STR) is very challenging due to the diversity of text instances and the complexity of scenes.
We propose automated STR (AutoSTR) to search data-dependent backbones to boost text recognition performance.
Experiments demonstrate that, by searching data-dependent backbones, AutoSTR can outperform the state-of-the-art approaches on standard benchmarks.
arXiv Detail & Related papers (2020-03-14T06:51:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.