Font Generation with Missing Impression Labels
- URL: http://arxiv.org/abs/2203.10348v1
- Date: Sat, 19 Mar 2022 16:02:54 GMT
- Title: Font Generation with Missing Impression Labels
- Authors: Seiya Matsuda, Akisato Kimura, Seiichi Uchida
- Abstract summary: This paper proposes a font generation model that is robust against missing impression labels.
Key ideas of the proposed method are (1)a co-occurrence-based missing label estimator and (2)an impression label space compressor.
We proved that the proposed model generates high-quality font images using multi-label data with missing labels through qualitative and quantitative evaluations.
- Score: 15.228202509283248
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Our goal is to generate fonts with specific impressions, by training a
generative adversarial network with a font dataset with impression labels. The
main difficulty is that font impression is ambiguous and the absence of an
impression label does not always mean that the font does not have the
impression. This paper proposes a font generation model that is robust against
missing impression labels. The key ideas of the proposed method are (1)a
co-occurrence-based missing label estimator and (2)an impression label space
compressor. The first is to interpolate missing impression labels based on the
co-occurrence of labels in the dataset and use them for training the model as
completed label conditions. The second is an encoder-decoder module to compress
the high-dimensional impression space into low-dimensional. We proved that the
proposed model generates high-quality font images using multi-label data with
missing labels through qualitative and quantitative evaluations.
Related papers
- Embedding Font Impression Word Tags Based on Co-occurrence [7.312180925669324]
Different font styles convey distinct impressions, indicating a close relationship between font shapes and word tags describing those impressions.<n>This paper proposes a novel embedding method for impression tags that leverages these shape-impression relationships.
arXiv Detail & Related papers (2025-08-26T09:02:17Z) - Sanitizing Manufacturing Dataset Labels Using Vision-Language Models [1.0819408603463427]
This paper introduces Vision-Language Sanitization and Refinement (VLSR), which is a vision-language-based framework for label sanitization and refinement.<n>The method embeds both images and their associated textual labels into a shared semantic space leveraging the CLIP vision-language model.<n> Experimental results demonstrate that the VLSR framework successfully identifies problematic labels and improves label consistency.
arXiv Detail & Related papers (2025-06-30T02:13:09Z) - Impression-CLIP: Contrastive Shape-Impression Embedding for Fonts [7.542892664684078]
We propose Impression-CLIP, which is a novel machine-learning model based on CLIP (Contrastive Language-Image Pre-training)
In our experiment, we perform cross-modal retrieval between fonts and impressions through co-embedding.
The results indicate that Impression-CLIP achieves better retrieval accuracy than the state-of-the-art method.
arXiv Detail & Related papers (2024-02-26T07:07:18Z) - Bridging the Gap between Model Explanations in Partially Annotated
Multi-label Classification [85.76130799062379]
We study how false negative labels affect the model's explanation.
We propose to boost the attribution scores of the model trained with partial labels to make its explanation resemble that of the model trained with full labels.
arXiv Detail & Related papers (2023-04-04T14:00:59Z) - Seq-UPS: Sequential Uncertainty-aware Pseudo-label Selection for
Semi-Supervised Text Recognition [21.583569162994277]
One of the most popular SSL approaches is pseudo-labeling (PL)
PL methods are severely degraded by noise and are prone to over-fitting to noisy labels.
We propose a pseudo-label generation and an uncertainty-based data selection framework for text recognition.
arXiv Detail & Related papers (2022-08-31T02:21:02Z) - Dual-Perspective Semantic-Aware Representation Blending for Multi-Label
Image Recognition with Partial Labels [70.36722026729859]
We propose a dual-perspective semantic-aware representation blending (DSRB) that blends multi-granularity category-specific semantic representation across different images.
The proposed DS consistently outperforms current state-of-the-art algorithms on all proportion label settings.
arXiv Detail & Related papers (2022-05-26T00:33:44Z) - Label Semantics for Few Shot Named Entity Recognition [68.01364012546402]
We study the problem of few shot learning for named entity recognition.
We leverage the semantic information in the names of the labels as a way of giving the model additional signal and enriched priors.
Our model learns to match the representations of named entities computed by the first encoder with label representations computed by the second encoder.
arXiv Detail & Related papers (2022-03-16T23:21:05Z) - Semantic-Aware Representation Blending for Multi-Label Image Recognition
with Partial Labels [86.17081952197788]
We propose to blend category-specific representation across different images to transfer information of known labels to complement unknown labels.
Experiments on the MS-COCO, Visual Genome, Pascal VOC 2007 datasets show that the proposed SARB framework obtains superior performance over current leading competitors.
arXiv Detail & Related papers (2022-03-04T07:56:16Z) - Inferring Prototypes for Multi-Label Few-Shot Image Classification with
Word Vector Guided Attention [45.6809084493491]
Multi-label few-shot image classification (ML-FSIC) is the task of assigning descriptive labels to previously unseen images.
In this paper we propose to use word embeddings as a form of prior knowledge about the meaning of the labels.
Our model can infer prototypes for unseen labels without the need for fine-tuning any model parameters.
arXiv Detail & Related papers (2021-12-02T07:59:11Z) - Field Extraction from Forms with Unlabeled Data [53.909807775291746]
We propose a novel framework to conduct field extraction from forms with unlabeled data.
We develop a rule-based method for mining noisy pseudo-labels from unlabeled forms.
arXiv Detail & Related papers (2021-10-08T17:50:12Z) - A Study on the Autoregressive and non-Autoregressive Multi-label
Learning [77.11075863067131]
We propose a self-attention based variational encoder-model to extract the label-label and label-feature dependencies jointly.
Our model can therefore be used to predict all labels in parallel while still including both label-label and label-feature dependencies.
arXiv Detail & Related papers (2020-12-03T05:41:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.