TNCR: Table Net Detection and Classification Dataset
- URL: http://arxiv.org/abs/2106.15322v1
- Date: Sat, 19 Jun 2021 10:48:58 GMT
- Title: TNCR: Table Net Detection and Classification Dataset
- Authors: Abdelrahman Abdallah, Alexander Berendeyev, Islam Nuradin, Daniyar
Nurseitov
- Abstract summary: TNCR dataset can be used for table detection in scanned document images and their classification into 5 different classes.
We have implemented state-of-the-art deep learning-based methods for table detection to create several strong baselines.
We have made TNCR open source in the hope of encouraging more deep learning approaches to table detection, classification, and structure recognition.
- Score: 62.997667081978825
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present TNCR, a new table dataset with varying image quality collected
from free websites. The TNCR dataset can be used for table detection in scanned
document images and their classification into 5 different classes. TNCR
contains 9428 high-quality labeled images. In this paper, we have implemented
state-of-the-art deep learning-based methods for table detection to create
several strong baselines. Cascade Mask R-CNN with ResNeXt-101-64x4d Backbone
Network achieves the highest performance compared to other methods with a
precision of 79.7%, recall of 89.8%, and f1 score of 84.4% on the TNCR dataset.
We have made TNCR open source in the hope of encouraging more deep learning
approaches to table detection, classification, and structure recognition. The
dataset and trained model checkpoints are available at
https://github.com/abdoelsayed2016/TNCR_Dataset.
Related papers
- Fuzzy Convolution Neural Networks for Tabular Data Classification [0.0]
Convolutional neural networks (CNNs) have attracted a great deal of attention due to their remarkable performance in various domains.
In this paper, we propose a novel framework fuzzy convolution neural network (FCNN) tailored specifically for tabular data.
arXiv Detail & Related papers (2024-06-04T20:33:35Z) - Feature Extraction and Classification from Planetary Science Datasets
enabled by Machine Learning [0.4091406230302996]
We present two examples of recent investigations, applying Machine Learning (ML) neural networks to image datasets from outer planet missions to achieve feature recognition.
We used a transfer learning approach, adding and training new layers to an industry-standard Mask R-CNN to recognize labeled blocks in a training dataset.
In a different application, we applied the Mask R-CNN to recognize clouds on Titan, again through updated training followed by testing against new data, with a precision of 95% over 369 images.
arXiv Detail & Related papers (2023-10-26T11:43:55Z) - One-Shot Learning for Periocular Recognition: Exploring the Effect of
Domain Adaptation and Data Bias on Deep Representations [59.17685450892182]
We investigate the behavior of deep representations in widely used CNN models under extreme data scarcity for One-Shot periocular recognition.
We improved state-of-the-art results that made use of networks trained with biometric datasets with millions of images.
Traditional algorithms like SIFT can outperform CNNs in situations with limited data.
arXiv Detail & Related papers (2023-07-11T09:10:16Z) - Text recognition on images using pre-trained CNN [2.191505742658975]
The recognition is trained by using Chars74K dataset and the best model results then tested on some samples of IIIT-5K-Dataset.
The research model has an accuracy of 97.94% for validation data, 98.16% for test data, and 95.62% for the test data from IIIT-5K-Dataset.
arXiv Detail & Related papers (2023-02-10T08:09:51Z) - RankDNN: Learning to Rank for Few-shot Learning [70.49494297554537]
This paper introduces a new few-shot learning pipeline that casts relevance ranking for image retrieval as binary ranking relation classification.
It provides a new perspective on few-shot learning and is complementary to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-28T13:59:31Z) - CNN Filter Learning from Drawn Markers for the Detection of Suggestive
Signs of COVID-19 in CT Images [58.720142291102135]
We propose a method that does not require either large annotated datasets or backpropagation to estimate the filters of a convolutional neural network (CNN)
For a few CT images, the user draws markers at representative normal and abnormal regions.
The method generates a feature extractor composed of a sequence of convolutional layers, whose kernels are specialized in enhancing regions similar to the marked ones.
arXiv Detail & Related papers (2021-11-16T15:03:42Z) - SCARF: Self-Supervised Contrastive Learning using Random Feature
Corruption [72.35532598131176]
We propose SCARF, a technique for contrastive learning, where views are formed by corrupting a random subset of features.
We show that SCARF complements existing strategies and outperforms alternatives like autoencoders.
arXiv Detail & Related papers (2021-06-29T08:08:33Z) - Tamil Vowel Recognition With Augmented MNIST-like Data Set [0.0]
We report the capability of the 60,000 grayscale, 28x28 pixel to build a 92% accuracy dataset.
We also report a top-1 classification accuracy of 70% and top-2 classification accuracy of 92% on handwritten vowels, for the same network.
arXiv Detail & Related papers (2020-06-09T19:17:30Z) - CascadeTabNet: An approach for end to end table detection and structure
recognition from image-based documents [4.199844472131922]
We present an improved deep learning-based end to end approach for solving both problems of table detection and structure recognition.
We propose CascadeTabNet: a Cascade mask Region-based CNN High-Resolution Network ( Cascade mask R-CNN HRNet) based model.
We attain the highest accuracy results on the ICDAR 2019 table structure recognition dataset.
arXiv Detail & Related papers (2020-04-27T08:12:48Z) - R-FCN: Object Detection via Region-based Fully Convolutional Networks [87.62557357527861]
We present region-based, fully convolutional networks for accurate and efficient object detection.
Our result is achieved at a test-time speed of 170ms per image, 2.5-20x faster than the Faster R-CNN counterpart.
arXiv Detail & Related papers (2016-05-20T15:50:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.