Interpretable Distance Metric Learning for Handwritten Chinese Character
Recognition
- URL: http://arxiv.org/abs/2103.09714v1
- Date: Wed, 17 Mar 2021 15:17:02 GMT
- Title: Interpretable Distance Metric Learning for Handwritten Chinese Character
Recognition
- Authors: Boxiang Dong, Aparna S. Varde, Danilo Stevanovic, Jiayin Wang, Liang
Zhao
- Abstract summary: We propose an interpretable distance metric learning approach for handwritten Chinese character recognition.
Our experimental results on a benchmark dataset demonstrate the superior efficiency, accuracy and interpretability of our proposed approach.
- Score: 8.233701182710035
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Handwriting recognition is of crucial importance to both Human Computer
Interaction (HCI) and paperwork digitization. In the general field of Optical
Character Recognition (OCR), handwritten Chinese character recognition faces
tremendous challenges due to the enormously large character sets and the
amazing diversity of writing styles. Learning an appropriate distance metric to
measure the difference between data inputs is the foundation of accurate
handwritten character recognition. Existing distance metric learning approaches
either produce unacceptable error rates, or provide little interpretability in
the results. In this paper, we propose an interpretable distance metric
learning approach for handwritten Chinese character recognition. The learned
metric is a linear combination of intelligible base metrics, and thus provides
meaningful insights to ordinary users. Our experimental results on a benchmark
dataset demonstrate the superior efficiency, accuracy and interpretability of
our proposed approach.
Related papers
- Self-Supervised Representation Learning with Spatial-Temporal Consistency for Sign Language Recognition [96.62264528407863]
We propose a self-supervised contrastive learning framework to excavate rich context via spatial-temporal consistency.
Inspired by the complementary property of motion and joint modalities, we first introduce first-order motion information into sign language modeling.
Our method is evaluated with extensive experiments on four public benchmarks, and achieves new state-of-the-art performance with a notable margin.
arXiv Detail & Related papers (2024-06-15T04:50:19Z) - Machine Translation Meta Evaluation through Translation Accuracy
Challenge Sets [92.38654521870444]
We introduce ACES, a contrastive challenge set spanning 146 language pairs.
This dataset aims to discover whether metrics can identify 68 translation accuracy errors.
We conduct a large-scale study by benchmarking ACES on 50 metrics submitted to the WMT 2022 and 2023 metrics shared tasks.
arXiv Detail & Related papers (2024-01-29T17:17:42Z) - Multiscale Feature Learning Using Co-Tuplet Loss for Offline Handwritten Signature Verification [0.0]
We introduce the MultiScale Signature feature learning Network (MS-SigNet) with the co-tuplet loss.
MS-SigNet learns both global and regional signature features from multiple spatial scales, enhancing feature discrimination.
We also present HanSig, a large-scale Chinese signature dataset to support robust system development for this language.
arXiv Detail & Related papers (2023-08-01T10:14:43Z) - A Dataset of Inertial Measurement Units for Handwritten English
Alphabets [16.74710649245842]
This paper presents an end-to-end methodology for collecting datasets to recognize handwritten English alphabets.
The IMUs are utilized to capture the dynamic movement patterns associated with handwriting, enabling more accurate recognition of alphabets.
arXiv Detail & Related papers (2023-07-05T17:54:36Z) - End-to-End Page-Level Assessment of Handwritten Text Recognition [69.55992406968495]
HTR systems increasingly face the end-to-end page-level transcription of a document.
Standard metrics do not take into account the inconsistencies that might appear.
We propose a two-fold evaluation, where the transcription accuracy and the RO goodness are considered separately.
arXiv Detail & Related papers (2023-01-14T15:43:07Z) - Reading and Writing: Discriminative and Generative Modeling for
Self-Supervised Text Recognition [101.60244147302197]
We introduce contrastive learning and masked image modeling to learn discrimination and generation of text images.
Our method outperforms previous self-supervised text recognition methods by 10.2%-20.2% on irregular scene text recognition datasets.
Our proposed text recognizer exceeds previous state-of-the-art text recognition methods by averagely 5.3% on 11 benchmarks, with similar model size.
arXiv Detail & Related papers (2022-07-01T03:50:26Z) - Transparent Human Evaluation for Image Captioning [70.03979566548823]
We develop a rubric-based human evaluation protocol for image captioning models.
We show that human-generated captions show substantially higher quality than machine-generated ones.
We hope that this work will promote a more transparent evaluation protocol for image captioning.
arXiv Detail & Related papers (2021-11-17T07:09:59Z) - Towards an IMU-based Pen Online Handwriting Recognizer [2.6707647984082357]
We present a online handwriting recognition system for word recognition based on inertial measurement units (IMUs)
This is obtained by means of a sensor-equipped pen that provides acceleration, angular velocity, and magnetic forces streamed via Bluetooth.
Our model combines convolutional and bidirectional LSTM networks, and is trained with the Connectionist Temporal Classification loss.
arXiv Detail & Related papers (2021-05-26T09:47:19Z) - Skeleton Based Sign Language Recognition Using Whole-body Keypoints [71.97020373520922]
Sign language is used by deaf or speech impaired people to communicate.
Skeleton-based recognition is becoming popular that it can be further ensembled with RGB-D based method to achieve state-of-the-art performance.
Inspired by the recent development of whole-body pose estimation citejin 2020whole, we propose recognizing sign language based on the whole-body key points and features.
arXiv Detail & Related papers (2021-03-16T03:38:17Z) - Spectral Graph-based Features for Recognition of Handwritten Characters:
A Case Study on Handwritten Devanagari Numerals [0.0]
We propose an approach that exploits the robust graph representation and spectral graph embedding concept to represent handwritten characters.
For corroboration of the efficacy of the proposed method, extensive experiments were carried out on the standard handwritten numeral Computer Vision Pattern Recognition, Unit of Indian Statistical Institute Kolkata dataset.
arXiv Detail & Related papers (2020-07-07T08:40:08Z) - Neural Computing for Online Arabic Handwriting Character Recognition
using Hard Stroke Features Mining [0.0]
An enhanced method of detecting the desired critical points from vertical and horizontal direction-length of handwriting stroke features of online Arabic script recognition is proposed.
A minimum feature set is extracted from these tokens for classification of characters using a multilayer perceptron with a back-propagation learning algorithm and modified sigmoid function-based activation function.
The proposed method achieves an average accuracy of 98.6% comparable in state of art character recognition techniques.
arXiv Detail & Related papers (2020-05-02T23:17:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.