On-Device Text Image Super Resolution
- URL: http://arxiv.org/abs/2011.10251v1
- Date: Fri, 20 Nov 2020 07:49:48 GMT
- Title: On-Device Text Image Super Resolution
- Authors: Dhruval Jain, Arun D Prabhu, Gopi Ramena, Manoj Goyal, Debi Prasanna
Mohanty, Sukumar Moharana, Naresh Purre
- Abstract summary: We present a novel deep neural network that reconstructs sharper character edges and thus boosts OCR confidence.
The proposed architecture not only achieves significant improvement in PSNR over bicubic upsampling but also runs with an average inference time of 11.7 ms per image.
We also achieve an OCR accuracy of 75.89% on the ICDAR 2015 TextSR dataset, where ground truth has an accuracy of 78.10%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recent research on super-resolution (SR) has witnessed major developments
with the advancements of deep convolutional neural networks. There is a need
for information extraction from scenic text images or even document images on
device, most of which are low-resolution (LR) images. Therefore, SR becomes an
essential pre-processing step as Bicubic Upsampling, which is conventionally
present in smartphones, performs poorly on LR images. To give the user more
control over his privacy, and to reduce the carbon footprint by reducing the
overhead of cloud computing and hours of GPU usage, executing SR models on the
edge is a necessity in the recent times. There are various challenges in
running and optimizing a model on resource-constrained platforms like
smartphones. In this paper, we present a novel deep neural network that
reconstructs sharper character edges and thus boosts OCR confidence. The
proposed architecture not only achieves significant improvement in PSNR over
bicubic upsampling on various benchmark datasets but also runs with an average
inference time of 11.7 ms per image. We have outperformed state-of-the-art on
the Text330 dataset. We also achieve an OCR accuracy of 75.89% on the ICDAR
2015 TextSR dataset, where ground truth has an accuracy of 78.10%.
Related papers
- Rethinking Image Super-Resolution from Training Data Perspectives [54.28824316574355]
We investigate the understudied effect of the training data used for image super-resolution (SR)
With this, we propose an automated image evaluation pipeline.
We find that datasets with (i) low compression artifacts, (ii) high within-image diversity as judged by the number of different objects, and (iii) a large number of images from ImageNet or PASS all positively affect SR performance.
arXiv Detail & Related papers (2024-09-01T16:25:04Z) - Compiler-Aware Neural Architecture Search for On-Mobile Real-time
Super-Resolution [48.13296296287587]
We propose a compiler-aware SR neural architecture search (NAS) framework that conducts depth search and per-layer width search with adaptive SR blocks.
We achieve real-time SR inference for implementing 720p resolution with competitive SR performance on GPU/DSP of mobile platforms.
arXiv Detail & Related papers (2022-07-25T23:59:19Z) - Real-Time Super-Resolution for Real-World Images on Mobile Devices [11.632812550056173]
Image Super-Resolution (ISR) aims at recovering High-Resolution (HR) images from the corresponding Low-Resolution (LR) counterparts.
Recent progress in ISR has been remarkable, but they are way too computationally intensive to be deployed on edge devices.
In this work, an approach for real-time ISR on mobile devices is presented, which is able to deal with a wide range of degradations in real-world scenarios.
arXiv Detail & Related papers (2022-06-03T18:44:53Z) - Hybrid Pixel-Unshuffled Network for Lightweight Image Super-Resolution [64.54162195322246]
Convolutional neural network (CNN) has achieved great success on image super-resolution (SR)
Most deep CNN-based SR models take massive computations to obtain high performance.
We propose a novel Hybrid Pixel-Unshuffled Network (HPUN) by introducing an efficient and effective downsampling module into the SR task.
arXiv Detail & Related papers (2022-03-16T20:10:41Z) - Improving Super-Resolution Performance using Meta-Attention Layers [17.870338228921327]
Convolutional Neural Networks (CNNs) have achieved impressive results across many super-resolution (SR) and image restoration tasks.
Ill-posed nature of SR can make it difficult to accurately super-resolve an image which has undergone multiple different degradations.
We introduce meta-attention, a mechanism which allows any SR CNN to exploit the information available in relevant degradation parameters.
arXiv Detail & Related papers (2021-10-27T09:20:21Z) - Achieving on-Mobile Real-Time Super-Resolution with Neural Architecture
and Pruning Search [64.80878113422824]
We propose an automatic search framework that derives sparse super-resolution (SR) models with high image quality while satisfying the real-time inference requirement.
With the proposed framework, we are the first to achieve real-time SR inference (with only tens of milliseconds per frame) for implementing 720p resolution with competitive image quality.
arXiv Detail & Related papers (2021-08-18T06:47:31Z) - Deep Burst Super-Resolution [165.90445859851448]
We propose a novel architecture for the burst super-resolution task.
Our network takes multiple noisy RAW images as input, and generates a denoised, super-resolved RGB image as output.
In order to enable training and evaluation on real-world data, we additionally introduce the BurstSR dataset.
arXiv Detail & Related papers (2021-01-26T18:57:21Z) - Boosting High-Level Vision with Joint Compression Artifacts Reduction
and Super-Resolution [10.960291115491504]
We generate an artifact-free high-resolution image from a low-resolution one compressed with an arbitrary quality factor.
A context-aware joint CAR and SR neural network (CAJNN) integrates both local and non-local features to solve CAR and SR in one-stage.
A deep reconstruction network is adopted to predict high quality and high-resolution images.
arXiv Detail & Related papers (2020-10-18T04:17:08Z) - Scene Text Image Super-Resolution in the Wild [112.90416737357141]
Low-resolution text images are often seen in natural scenes such as documents captured by mobile phones.
Previous single image super-resolution (SISR) methods are trained on synthetic low-resolution images.
We pro-pose a real scene text SR dataset, termed TextZoom.
It contains paired real low-resolution and high-resolution images captured by cameras with different focal length in the wild.
arXiv Detail & Related papers (2020-05-07T09:18:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.