Full Reference Screen Content Image Quality Assessment by Fusing
Multi-level Structure Similarity
- URL: http://arxiv.org/abs/2008.05396v1
- Date: Fri, 7 Aug 2020 10:20:25 GMT
- Title: Full Reference Screen Content Image Quality Assessment by Fusing
Multi-level Structure Similarity
- Authors: Chenglizhao Chen, Hongmeng Zhao, Huan Yang, Chong Peng, Teng Yu
- Abstract summary: This paper advocates a novel solution to measure structure similarity "globally" from the perspective of sparse representation.
To perform multi-level quality assessment in accordance with the real HVS, the above-mentioned global metric will be integrated with the conventional local ones by resorting to the newly devised selective deep fusion network.
- Score: 27.971146869941112
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The screen content images (SCIs) usually comprise various content types with
sharp edges, in which the artifacts or distortions can be well sensed by the
vanilla structure similarity measurement in a full reference manner.
Nonetheless, almost all of the current SOTA structure similarity metrics are
"locally" formulated in a single-level manner, while the true human visual
system (HVS) follows the multi-level manner, and such mismatch could eventually
prevent these metrics from achieving trustworthy quality assessment. To
ameliorate, this paper advocates a novel solution to measure structure
similarity "globally" from the perspective of sparse representation. To perform
multi-level quality assessment in accordance with the real HVS, the
above-mentioned global metric will be integrated with the conventional local
ones by resorting to the newly devised selective deep fusion network. To
validate its efficacy and effectiveness, we have compared our method with 12
SOTA methods over two widely-used large-scale public SCI datasets, and the
quantitative results indicate that our method yields significantly higher
consistency with subjective quality score than the currently leading works.
Both the source code and data are also publicly available to gain widespread
acceptance and facilitate new advancement and its validation.
Related papers
- A Quality-Guided Mixture of Score-Fusion Experts Framework for Human Recognition [14.242680363313148]
Whole-body biometric recognition is a challenging task that integrates various biometric modalities.<n>We present textbfQuality-guided textbfMixture of score-fusion textbfExperts (QME)<n>We introduce a novel pseudo-quality loss for quality estimation with a modality-specific Quality Estimator (QE) and a score triplet loss to improve the metric performance.
arXiv Detail & Related papers (2025-07-31T18:00:01Z) - CMET: Clustering guided METric for quantifying embedding quality [0.0]
Clustering guided METric (CMET) is a metric for quantifying embedding quality.<n>CMET consists of two scores, viz., CMET_L and CMET_G, that measure the degree of local and global shape preservation capability.<n>Results reflect the favorable performance of CMET against the state-of-the-art methods.
arXiv Detail & Related papers (2025-07-07T10:02:34Z) - Leveraging Vision-Language Models to Select Trustworthy Super-Resolution Samples Generated by Diffusion Models [0.026861992804651083]
This paper introduces a robust framework for identifying the most trustworthy SR sample from a diffusion-generated set.<n>We propose a novel Trustworthiness Score (TWS) a hybrid metric that quantifies SR reliability based on semantic similarity.<n>By aligning outputs with human expectations and semantic correctness, this work sets a new benchmark for trustworthiness in generative SR.
arXiv Detail & Related papers (2025-06-25T21:00:44Z) - No-Reference Image Quality Assessment with Global-Local Progressive Integration and Semantic-Aligned Quality Transfer [6.095342999639137]
We develop a dual-measurement framework that combines vision Transformer (ViT)-based global feature extractor and convolutional neural networks (CNNs)-based local feature extractor.
We introduce a semantic-aligned quality transfer method that extends the training data by automatically labeling the quality scores of diverse image content with subjective opinion scores.
arXiv Detail & Related papers (2024-08-07T16:34:32Z) - Siamese Transformer Networks for Few-shot Image Classification [9.55588609556447]
Humans exhibit remarkable proficiency in visual classification tasks, accurately recognizing and classifying new images with minimal examples.
Existing few-shot image classification methods often emphasize either global features or local features, with few studies considering the integration of both.
We propose a novel approach based on the Siamese Transformer Network (STN)
Our strategy effectively harnesses the potential of global and local features in few-shot image classification, circumventing the need for complex feature adaptation modules.
arXiv Detail & Related papers (2024-07-16T14:27:23Z) - FlowCon: Out-of-Distribution Detection using Flow-Based Contrastive Learning [0.0]
We introduce textitFlowCon, a new density-based OOD detection technique.
Our main innovation lies in efficiently combining the properties of normalizing flow with supervised contrastive learning.
Empirical evaluation shows the enhanced performance of our method across common vision datasets.
arXiv Detail & Related papers (2024-07-03T20:33:56Z) - Efficient Multi-View Graph Clustering with Local and Global Structure
Preservation [59.49018175496533]
We propose a novel anchor-based multi-view graph clustering framework termed Efficient Multi-View Graph Clustering with Local and Global Structure Preservation (EMVGC-LG)
Specifically, EMVGC-LG jointly optimize anchor construction and graph learning to enhance the clustering quality.
In addition, EMVGC-LG inherits the linear complexity of existing AMVGC methods respecting the sample number.
arXiv Detail & Related papers (2023-08-31T12:12:30Z) - DealMVC: Dual Contrastive Calibration for Multi-view Clustering [78.54355167448614]
We propose a novel Dual contrastive calibration network for Multi-View Clustering (DealMVC)
We first design a fusion mechanism to obtain a global cross-view feature. Then, a global contrastive calibration loss is proposed by aligning the view feature similarity graph and the high-confidence pseudo-label graph.
During the training procedure, the interacted cross-view feature is jointly optimized at both local and global levels.
arXiv Detail & Related papers (2023-08-17T14:14:28Z) - GREAT Score: Global Robustness Evaluation of Adversarial Perturbation using Generative Models [60.48306899271866]
We present a new framework, called GREAT Score, for global robustness evaluation of adversarial perturbation using generative models.
We show high correlation and significantly reduced cost of GREAT Score when compared to the attack-based model ranking on RobustBench.
GREAT Score can be used for remote auditing of privacy-sensitive black-box models.
arXiv Detail & Related papers (2023-04-19T14:58:27Z) - TRUE: Re-evaluating Factual Consistency Evaluation [29.888885917330327]
We introduce TRUE: a comprehensive study of factual consistency metrics on a standardized collection of existing texts from diverse tasks.
Our standardization enables an example-level meta-evaluation protocol that is more actionable and interpretable than previously reported correlations.
Across diverse state-of-the-art metrics and 11 datasets we find that large-scale NLI and question generation-and-answering-based approaches achieve strong and complementary results.
arXiv Detail & Related papers (2022-04-11T10:14:35Z) - Quality-Aware Multimodal Biometric Recognition [30.322429033099688]
We develop a quality-aware framework for fusing representations of input modalities by weighting their importance using quality scores estimated in a weakly-supervised fashion.
This framework utilizes two fusion blocks, each represented by a set of quality-aware and aggregation networks.
We evaluate the performance by considering three multimodal datasets consisting of face, iris, and fingerprint modalities.
arXiv Detail & Related papers (2021-12-10T20:48:55Z) - Open-Set Recognition: A Good Closed-Set Classifier is All You Need [146.6814176602689]
We show that the ability of a classifier to make the 'none-of-above' decision is highly correlated with its accuracy on the closed-set classes.
We use this correlation to boost the performance of the cross-entropy OSR 'baseline' by improving its closed-set accuracy.
We also construct new benchmarks which better respect the task of detecting semantic novelty.
arXiv Detail & Related papers (2021-10-12T17:58:59Z) - Semi-Supervised Domain Generalization with Stochastic StyleMatch [90.98288822165482]
In real-world applications, we might have only a few labels available from each source domain due to high annotation cost.
In this work, we investigate semi-supervised domain generalization, a more realistic and practical setting.
Our proposed approach, StyleMatch, is inspired by FixMatch, a state-of-the-art semi-supervised learning method based on pseudo-labeling.
arXiv Detail & Related papers (2021-06-01T16:00:08Z) - Trusted Multi-View Classification [76.73585034192894]
We propose a novel multi-view classification method, termed trusted multi-view classification.
It provides a new paradigm for multi-view learning by dynamically integrating different views at an evidence level.
The proposed algorithm jointly utilizes multiple views to promote both classification reliability and robustness.
arXiv Detail & Related papers (2021-02-03T13:30:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.