DSS: Synthesizing long Digital Ink using Data augmentation, Style
encoding and Split generation
- URL: http://arxiv.org/abs/2311.17786v1
- Date: Wed, 29 Nov 2023 16:33:19 GMT
- Title: DSS: Synthesizing long Digital Ink using Data augmentation, Style
encoding and Split generation
- Authors: Aleksandr Timofeev, Anastasiia Fadeeva, Andrei Afonin, Claudiu Musat,
Andrii Maksai
- Abstract summary: We show that the commonly used models for this task fail to generalize to long-form data.
These methods use contrastive learning technique and are tailored specifically for the handwriting domain.
- Score: 47.90135553071684
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As text generative models can give increasingly long answers, we tackle the
problem of synthesizing long text in digital ink. We show that the commonly
used models for this task fail to generalize to long-form data and how this
problem can be solved by augmenting the training data, changing the model
architecture and the inference procedure. These methods use contrastive
learning technique and are tailored specifically for the handwriting domain.
They can be applied to any encoder-decoder model that works with digital ink.
We demonstrate that our method reduces the character error rate on long-form
English data by half compared to baseline RNN and by 16% compared to the
previous approach that aims at addressing the same problem. We show that all
three parts of the method improve recognizability of generated inks. In
addition, we evaluate synthesized data in a human study and find that people
perceive most of generated data as real.
Related papers
- Arbitrary-Length Generalization for Addition in a Tiny Transformer [55.2480439325792]
This paper introduces a novel training methodology that enables a Transformer model to generalize the addition of two-digit numbers to numbers with unseen lengths of digits.
The proposed approach employs an autoregressive generation technique, processing from right to left, which mimics a common manual method for adding large numbers.
arXiv Detail & Related papers (2024-05-31T03:01:16Z) - InkSight: Offline-to-Online Handwriting Conversion by Learning to Read
and Write [7.827729986700937]
InkSight aims to empower physical note-takers to effortlessly convert their work (offline handwriting) to digital ink (online handwriting)
Our approach combines reading and writing priors, allowing training a model in the absence of large amounts of paired samples.
Our human evaluation reveals that 87% of the samples produced by our model on the challenging HierText dataset are considered as a valid tracing of the input image.
arXiv Detail & Related papers (2024-02-08T16:41:41Z) - Deep Learning-Driven Approach for Handwritten Chinese Character Classification [0.0]
Handwritten character recognition is a challenging problem for machine learning researchers.
With numerous unique character classes present, some data, such as Logographic Scripts or Sino-Korean character sequences, bring new complications to the HCR problem.
This paper proposes a highly scalable approach for detailed character image classification by introducing the model architecture, data preprocessing steps, and testing design instructions.
arXiv Detail & Related papers (2024-01-30T15:29:32Z) - GEC-DePenD: Non-Autoregressive Grammatical Error Correction with
Decoupled Permutation and Decoding [52.14832976759585]
Grammatical error correction (GEC) is an important NLP task that is usually solved with autoregressive sequence-to-sequence models.
We propose a novel non-autoregressive approach to GEC that decouples the architecture into a permutation network.
We show that the resulting network improves over previously known non-autoregressive methods for GEC.
arXiv Detail & Related papers (2023-11-14T14:24:36Z) - Critic-Driven Decoding for Mitigating Hallucinations in Data-to-text
Generation [5.304395026626743]
Hallucination of text ungrounded in the input is a well-known problem in neural data-to-text generation.
We propose a new way to mitigate hallucinations by combining the probabilistic output of a generator language model with the output of a special "text critic"
Our method does not need any changes to the underlying LM's architecture or training procedure.
arXiv Detail & Related papers (2023-10-25T20:05:07Z) - Optimizing Factual Accuracy in Text Generation through Dynamic Knowledge
Selection [71.20871905457174]
Language models (LMs) have revolutionized the way we interact with information, but they often generate nonfactual text.
Previous methods use external knowledge as references for text generation to enhance factuality but often struggle with the knowledge mix-up of irrelevant references.
We present DKGen, which divide the text generation process into an iterative process.
arXiv Detail & Related papers (2023-08-30T02:22:40Z) - Autoencoding for the 'Good Dictionary' of eigen pairs of the Koopman
Operator [0.0]
This paper proposes using deep autoencoders, a type of deep learning technique, to perform non-linear geometric transformations on raw data before computing Koopman eigen vectors.
To handle high-dimensional time series data, Takens's time delay embedding is presented as a pre-processing technique.
arXiv Detail & Related papers (2023-06-08T14:21:01Z) - Sampling and Ranking for Digital Ink Generation on a tight computational
budget [69.15275423815461]
We study ways to maximize the quality of the output of a trained digital ink generative model.
We use and compare the effect of multiple sampling and ranking techniques, in the first ablation study of its kind in the digital ink domain.
arXiv Detail & Related papers (2023-06-02T09:55:15Z) - Lexically Aware Semi-Supervised Learning for OCR Post-Correction [90.54336622024299]
Much of the existing linguistic data in many languages of the world is locked away in non-digitized books and documents.
Previous work has demonstrated the utility of neural post-correction methods on recognition of less-well-resourced languages.
We present a semi-supervised learning method that makes it possible to utilize raw images to improve performance.
arXiv Detail & Related papers (2021-11-04T04:39:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.