Tamil Vowel Recognition With Augmented MNIST-like Data Set
- URL: http://arxiv.org/abs/2006.08367v2
- Date: Tue, 16 Jun 2020 19:20:09 GMT
- Title: Tamil Vowel Recognition With Augmented MNIST-like Data Set
- Authors: Muthiah Annamalai
- Abstract summary: We report the capability of the 60,000 grayscale, 28x28 pixel to build a 92% accuracy dataset.
We also report a top-1 classification accuracy of 70% and top-2 classification accuracy of 92% on handwritten vowels, for the same network.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We report generation of a MNIST [4] compatible data set [1] for Tamil vowels
to enable building a classification DNN or other such ML/AI deep learning [2]
models for Tamil OCR/Handwriting applications. We report the capability of the
60,000 grayscale, 28x28 pixel dataset to build a 92% accuracy (training) and
82% cross-validation 4-layer CNN, with 100,000+ parameters, in TensorFlow. We
also report a top-1 classification accuracy of 70% and top-2 classification
accuracy of 92% on handwritten vowels showing, for the same network.
Related papers
- Text Quality-Based Pruning for Efficient Training of Language Models [66.66259229732121]
We propose a novel method for numerically evaluating text quality in large unlabelled NLP datasets.
By proposing the text quality metric, the paper establishes a framework to identify and eliminate low-quality text instances.
Experimental results over multiple models and datasets demonstrate the efficacy of this approach.
arXiv Detail & Related papers (2024-04-26T18:01:25Z) - Text recognition on images using pre-trained CNN [2.191505742658975]
The recognition is trained by using Chars74K dataset and the best model results then tested on some samples of IIIT-5K-Dataset.
The research model has an accuracy of 97.94% for validation data, 98.16% for test data, and 95.62% for the test data from IIIT-5K-Dataset.
arXiv Detail & Related papers (2023-02-10T08:09:51Z) - Deep Multi-Scale Representation Learning with Attention for Automatic
Modulation Classification [11.32380278232938]
We find some experienced improvements by using large kernel size for convolutional deep convolution neural network based AMC.
We propose a multi-scale feature network with large kernel size and SE mechanism (SE-MSFN) in this paper.
SE-MSFN achieves state-of-the-art classification performance on the public well-known RADIOML 2018.01A dataset.
arXiv Detail & Related papers (2022-08-31T07:26:09Z) - Facilitated machine learning for image-based fruit quality assessment in
developing countries [68.8204255655161]
Automated image classification is a common task for supervised machine learning in food science.
We propose an alternative method based on pre-trained vision transformers (ViTs)
It can be easily implemented with limited resources on a standard device.
arXiv Detail & Related papers (2022-07-10T19:52:20Z) - Alexa Teacher Model: Pretraining and Distilling Multi-Billion-Parameter
Encoders for Natural Language Understanding Systems [63.713297451300086]
We present results from a large-scale experiment on pretraining encoders with non-embedding parameter counts ranging from 700M to 9.3B.
Their subsequent distillation into smaller models ranging from 17M-170M parameters, and their application to the Natural Language Understanding (NLU) component of a virtual assistant system.
arXiv Detail & Related papers (2022-06-15T20:44:23Z) - OneAligner: Zero-shot Cross-lingual Transfer with One Rich-Resource
Language Pair for Low-Resource Sentence Retrieval [91.76575626229824]
We present OneAligner, an alignment model specially designed for sentence retrieval tasks.
When trained with all language pairs of a large-scale parallel multilingual corpus (OPUS-100), this model achieves the state-of-the-art result.
We conclude through empirical results and analyses that the performance of the sentence alignment task depends mostly on the monolingual and parallel data size.
arXiv Detail & Related papers (2022-05-17T19:52:42Z) - Data and knowledge-driven approaches for multilingual training to
improve the performance of speech recognition systems of Indian languages [0.0]
We propose data and knowledge-driven approaches for multilingual training of the automated speech recognition system for a target language.
In phone/senone mapping, deep neural network (DNN) learns to map senones or phones from one language to the others.
In the other approach, we model the acoustic information for all the languages simultaneously.
arXiv Detail & Related papers (2022-01-24T07:17:17Z) - Exploiting Hybrid Models of Tensor-Train Networks for Spoken Command
Recognition [9.262289183808035]
This work aims to design a low complexity spoken command recognition (SCR) system.
We exploit a deep hybrid architecture of a tensor-train (TT) network to build an end-to-end SRC pipeline.
Our proposed CNN+(TT-DNN) model attains a competitive accuracy of 96.31% with 4 times fewer model parameters than the CNN model.
arXiv Detail & Related papers (2022-01-11T05:57:38Z) - TNCR: Table Net Detection and Classification Dataset [62.997667081978825]
TNCR dataset can be used for table detection in scanned document images and their classification into 5 different classes.
We have implemented state-of-the-art deep learning-based methods for table detection to create several strong baselines.
We have made TNCR open source in the hope of encouraging more deep learning approaches to table detection, classification, and structure recognition.
arXiv Detail & Related papers (2021-06-19T10:48:58Z) - Fusion of CNNs and statistical indicators to improve image
classification [65.51757376525798]
Convolutional Networks have dominated the field of computer vision for the last ten years.
Main strategy to prolong this trend relies on further upscaling networks in size.
We hypothesise that adding heterogeneous sources of information may be more cost-effective to a CNN than building a bigger network.
arXiv Detail & Related papers (2020-12-20T23:24:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.