Enhance to Read Better: An Improved Generative Adversarial Network for
Handwritten Document Image Enhancement
- URL: http://arxiv.org/abs/2105.12710v1
- Date: Wed, 26 May 2021 17:44:45 GMT
- Title: Enhance to Read Better: An Improved Generative Adversarial Network for
Handwritten Document Image Enhancement
- Authors: Sana Khamekhem Jemni and Mohamed Ali Souibgui and Yousri Kessentini
and Alicia Forn\'es
- Abstract summary: We propose an end to end architecture based on Generative Adversarial Networks (GANs) to recover degraded documents into a clean and readable form.
To the best of our knowledge, this is the first work to use the text information while binarizing handwritten documents.
We outperform the state of the art in H-DIBCO 2018 challenge, after fine tuning our pre-trained model with synthetically degraded Latin handwritten images.
- Score: 1.7491858164568674
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Handwritten document images can be highly affected by degradation for
different reasons: Paper ageing, daily-life scenarios (wrinkles, dust, etc.),
bad scanning process and so on. These artifacts raise many readability issues
for current Handwritten Text Recognition (HTR) algorithms and severely devalue
their efficiency. In this paper, we propose an end to end architecture based on
Generative Adversarial Networks (GANs) to recover the degraded documents into a
clean and readable form. Unlike the most well-known document binarization
methods, which try to improve the visual quality of the degraded document, the
proposed architecture integrates a handwritten text recognizer that promotes
the generated document image to be more readable. To the best of our knowledge,
this is the first work to use the text information while binarizing handwritten
documents. Extensive experiments conducted on degraded Arabic and Latin
handwritten documents demonstrate the usefulness of integrating the recognizer
within the GAN architecture, which improves both the visual quality and the
readability of the degraded document images. Moreover, we outperform the state
of the art in H-DIBCO 2018 challenge, after fine tuning our pre-trained model
with synthetically degraded Latin handwritten images, on this task.
Related papers
- DocMAE: Document Image Rectification via Self-supervised Representation
Learning [144.44748607192147]
We present DocMAE, a novel self-supervised framework for document image rectification.
We first mask random patches of the background-excluded document images and then reconstruct the missing pixels.
With such a self-supervised learning approach, the network is encouraged to learn the intrinsic structure of deformed documents.
arXiv Detail & Related papers (2023-04-20T14:27:15Z) - Deep Unrestricted Document Image Rectification [110.61517455253308]
We present DocTr++, a novel unified framework for document image rectification.
We upgrade the original architecture by adopting a hierarchical encoder-decoder structure for multi-scale representation extraction and parsing.
We contribute a real-world test set and metrics applicable for evaluating the rectification quality.
arXiv Detail & Related papers (2023-04-18T08:00:54Z) - EraseNet: A Recurrent Residual Network for Supervised Document Cleaning [0.0]
This paper introduces a supervised approach for cleaning dirty documents using a new fully convolutional auto-encoder architecture.
The experiments in this paper have shown promising results as the model is able to learn a variety of ordinary as well as unusual noises and rectify them efficiently.
arXiv Detail & Related papers (2022-10-03T04:23:25Z) - Boosting Modern and Historical Handwritten Text Recognition with
Deformable Convolutions [52.250269529057014]
Handwritten Text Recognition (HTR) in free-volution pages is a challenging image understanding task.
We propose to adopt deformable convolutions, which can deform depending on the input at hand and better adapt to the geometric variations of the text.
arXiv Detail & Related papers (2022-08-17T06:55:54Z) - Open Set Classification of Untranscribed Handwritten Documents [56.0167902098419]
Huge amounts of digital page images of important manuscripts are preserved in archives worldwide.
The class or typology'' of a document is perhaps the most important tag to be included in the metadata.
The technical problem is one of automatic classification of documents, each consisting of a set of untranscribed handwritten text images.
arXiv Detail & Related papers (2022-06-20T20:43:50Z) - DocScanner: Robust Document Image Rectification with Progressive
Learning [162.03694280524084]
This work presents DocScanner, a new deep network architecture for document image rectification.
DocScanner maintains a single estimate of the rectified image, which is progressively corrected with a recurrent architecture.
The iterative refinements make DocScanner converge to a robust and superior performance, and the lightweight recurrent architecture ensures the running efficiency.
arXiv Detail & Related papers (2021-10-28T09:15:02Z) - SmartPatch: Improving Handwritten Word Imitation with Patch
Discriminators [67.54204685189255]
We propose SmartPatch, a new technique increasing the performance of current state-of-the-art methods.
We combine the well-known patch loss with information gathered from the parallel trained handwritten text recognition system.
This leads to a more enhanced local discriminator and results in more realistic and higher-quality generated handwritten words.
arXiv Detail & Related papers (2021-05-21T18:34:21Z) - DE-GAN: A Conditional Generative Adversarial Network for Document
Enhancement [4.073826298938431]
We propose an end-to-end framework named Document Enhancement Geneversarative Adrial Networks (DE-GAN) to restore severely degraded document images.
We demonstrate that, in different tasks (document clean up, binarization, deblurring and watermark removal), DE-GAN can produce an enhanced version of the degraded document with a high quality.
arXiv Detail & Related papers (2020-10-17T10:54:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.