Perception and Semantic Aware Regularization for Sequential Confidence
Calibration
- URL: http://arxiv.org/abs/2305.19498v1
- Date: Wed, 31 May 2023 02:16:29 GMT
- Title: Perception and Semantic Aware Regularization for Sequential Confidence
Calibration
- Authors: Zhenghua Peng, Yu Luo, Tianshui Chen, Keke Xu, Shuangping Huang
- Abstract summary: We propose a Perception and Semantic aware Sequence Regularization framework.
We introduce a semantic context-free recognition and a language model to acquire similar sequences with high perceptive similarities and semantic correlation.
Experiments on canonical sequence recognition tasks, including scene text and speech recognition, demonstrate that our method sets novel state-of-the-art results.
- Score: 12.265757315192497
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep sequence recognition (DSR) models receive increasing attention due to
their superior application to various applications. Most DSR models use merely
the target sequences as supervision without considering other related
sequences, leading to over-confidence in their predictions. The DSR models
trained with label smoothing regularize labels by equally and independently
smoothing each token, reallocating a small value to other tokens for mitigating
overconfidence. However, they do not consider tokens/sequences correlations
that may provide more effective information to regularize training and thus
lead to sub-optimal performance. In this work, we find tokens/sequences with
high perception and semantic correlations with the target ones contain more
correlated and effective information and thus facilitate more effective
regularization. To this end, we propose a Perception and Semantic aware
Sequence Regularization framework, which explore perceptively and semantically
correlated tokens/sequences as regularization. Specifically, we introduce a
semantic context-free recognition and a language model to acquire similar
sequences with high perceptive similarities and semantic correlation,
respectively. Moreover, over-confidence degree varies across samples according
to their difficulties. Thus, we further design an adaptive calibration
intensity module to compute a difficulty score for each samples to obtain
finer-grained regularization. Extensive experiments on canonical sequence
recognition tasks, including scene text and speech recognition, demonstrate
that our method sets novel state-of-the-art results. Code is available at
https://github.com/husterpzh/PSSR.
Related papers
- Dynamic Correlation Learning and Regularization for Multi-Label Confidence Calibration [60.95748658638956]
This paper introduces the Multi-Label Confidence task, aiming to provide well-calibrated confidence scores in multi-label scenarios.
Existing single-label calibration methods fail to account for category correlations, which are crucial for addressing semantic confusion.
We propose the Dynamic Correlation Learning and Regularization algorithm, which leverages multi-grained semantic correlations to better model semantic confusion.
arXiv Detail & Related papers (2024-07-09T13:26:21Z) - Noisy Correspondence Learning with Self-Reinforcing Errors Mitigation [63.180725016463974]
Cross-modal retrieval relies on well-matched large-scale datasets that are laborious in practice.
We introduce a novel noisy correspondence learning framework, namely textbfSelf-textbfReinforcing textbfErrors textbfMitigation (SREM)
arXiv Detail & Related papers (2023-12-27T09:03:43Z) - RankMatch: A Novel Approach to Semi-Supervised Label Distribution
Learning Leveraging Inter-label Correlations [52.549807652527306]
This paper introduces RankMatch, an innovative approach for Semi-Supervised Label Distribution Learning (SSLDL)
RankMatch effectively utilizes a small number of labeled examples in conjunction with a larger quantity of unlabeled data.
We establish a theoretical generalization bound for RankMatch, and through extensive experiments, demonstrate its superiority in performance against existing SSLDL methods.
arXiv Detail & Related papers (2023-12-11T12:47:29Z) - Enhancing Few-shot NER with Prompt Ordering based Data Augmentation [59.69108119752584]
We propose a Prompt Ordering based Data Augmentation (PODA) method to improve the training of unified autoregressive generation frameworks.
Experimental results on three public NER datasets and further analyses demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-05-19T16:25:43Z) - Context-Aware Selective Label Smoothing for Calibrating Sequence
Recognition Model [16.7796720078021]
We propose a Context-Aware Selective Label Smoothing (CASLS) method for calibrating sequential data.
Results on sequence recognition tasks, including scene text recognition and speech recognition, demonstrate that our method can achieve the state-of-the-art performance.
arXiv Detail & Related papers (2023-03-13T09:27:52Z) - How Does Beam Search improve Span-Level Confidence Estimation in
Generative Sequence Labeling? [11.481435098152893]
This paper aims to provide some empirical insights on estimating model confidence for generative sequence labeling.
As verified over six public datasets, we show that our proposed approach significantly reduces calibration errors of the predictions of a generative sequence labeling model.
arXiv Detail & Related papers (2022-12-21T05:01:01Z) - Neighbour Consistency Guided Pseudo-Label Refinement for Unsupervised
Person Re-Identification [80.98291772215154]
Unsupervised person re-identification (ReID) aims at learning discriminative identity features for person retrieval without any annotations.
Recent advances accomplish this task by leveraging clustering-based pseudo labels.
We propose a Neighbour Consistency guided Pseudo Label Refinement framework.
arXiv Detail & Related papers (2022-11-30T09:39:57Z) - Semi-Supervised Learning of Semantic Correspondence with Pseudo-Labels [26.542718087103665]
SemiMatch is a semi-supervised solution for establishing dense correspondences across semantically similar images.
Our framework generates the pseudo-labels using the model's prediction itself between source and weakly-augmented target, and uses pseudo-labels to learn the model again between source and strongly-augmented target.
In experiments, SemiMatch achieves state-of-the-art performance on various benchmarks, especially on PF-Willow by a large margin.
arXiv Detail & Related papers (2022-03-30T03:52:50Z) - Bi-Granularity Contrastive Learning for Post-Training in Few-Shot Scene [10.822477939237459]
We propose contrastive masked language modeling (CMLM) for post-training to integrate both token-level and sequence-level contrastive learnings.
CMLM surpasses several recent post-training methods in few-shot settings without the need for data augmentation.
arXiv Detail & Related papers (2021-06-04T08:17:48Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.