InfoMetIC: An Informative Metric for Reference-free Image Caption
Evaluation
- URL: http://arxiv.org/abs/2305.06002v1
- Date: Wed, 10 May 2023 09:22:44 GMT
- Title: InfoMetIC: An Informative Metric for Reference-free Image Caption
Evaluation
- Authors: Anwen Hu, Shizhe Chen, Liang Zhang, Qin Jin
- Abstract summary: We propose an Informative Metric for Reference-free Image Caption evaluation (InfoMetIC)
Given an image and a caption, InfoMetIC is able to report incorrect words and unmentioned image regions at fine-grained level.
We also construct a token-level evaluation dataset and demonstrate the effectiveness of InfoMetIC in fine-grained evaluation.
- Score: 69.1642316502563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automatic image captioning evaluation is critical for benchmarking and
promoting advances in image captioning research. Existing metrics only provide
a single score to measure caption qualities, which are less explainable and
informative. Instead, we humans can easily identify the problems of captions in
details, e.g., which words are inaccurate and which salient objects are not
described, and then rate the caption quality. To support such informative
feedback, we propose an Informative Metric for Reference-free Image Caption
evaluation (InfoMetIC). Given an image and a caption, InfoMetIC is able to
report incorrect words and unmentioned image regions at fine-grained level, and
also provide a text precision score, a vision recall score and an overall
quality score at coarse-grained level. The coarse-grained score of InfoMetIC
achieves significantly better correlation with human judgements than existing
metrics on multiple benchmarks. We also construct a token-level evaluation
dataset and demonstrate the effectiveness of InfoMetIC in fine-grained
evaluation. Our code and datasets are publicly available at
https://github.com/HAWLYQ/InfoMetIC.
Related papers
- BRIDGE: Bridging Gaps in Image Captioning Evaluation with Stronger Visual Cues [47.213906345208315]
We propose BRIDGE, a new learnable and reference-free image captioning metric.
Our proposal achieves state-of-the-art results compared to existing reference-free evaluation scores.
arXiv Detail & Related papers (2024-07-29T18:00:17Z) - FLEUR: An Explainable Reference-Free Evaluation Metric for Image Captioning Using a Large Multimodal Model [5.330266804358638]
We propose FLEUR, an explainable reference-free metric to introduce explainability into image captioning evaluation metrics.
By leveraging a large multimodal model, FLEUR can evaluate the caption against the image without the need for reference captions.
FLEUR achieves high correlations with human judgment across various image captioning evaluation benchmarks.
arXiv Detail & Related papers (2024-06-10T03:57:39Z) - Transparent Human Evaluation for Image Captioning [70.03979566548823]
We develop a rubric-based human evaluation protocol for image captioning models.
We show that human-generated captions show substantially higher quality than machine-generated ones.
We hope that this work will promote a more transparent evaluation protocol for image captioning.
arXiv Detail & Related papers (2021-11-17T07:09:59Z) - EMScore: Evaluating Video Captioning via Coarse-Grained and Fine-Grained
Embedding Matching [90.98122161162644]
Current metrics for video captioning are mostly based on the text-level comparison between reference and candidate captions.
We propose EMScore (Embedding Matching-based score), a novel reference-free metric for video captioning.
We exploit a well pre-trained vision-language model to extract visual and linguistic embeddings for computing EMScore.
arXiv Detail & Related papers (2021-11-17T06:02:43Z) - COSMic: A Coherence-Aware Generation Metric for Image Descriptions [27.41088864449921]
Image metrics have struggled to give accurate learned estimates of the semantic and pragmatic success of text evaluation models.
We present the first learned generation metric for evaluating output captions.
We demonstrate a higher out-efficient for our proposed metric the human judgments for the results of a number of state-of-the-art caption models when compared to several other metrics such as BLEURT and BERT.
arXiv Detail & Related papers (2021-09-11T13:43:36Z) - Contrastive Semantic Similarity Learning for Image Captioning Evaluation
with Intrinsic Auto-encoder [52.42057181754076]
Motivated by the auto-encoder mechanism and contrastive representation learning advances, we propose a learning-based metric for image captioning.
We develop three progressive model structures to learn the sentence level representations.
Experiment results show that our proposed method can align well with the scores generated from other contemporary metrics.
arXiv Detail & Related papers (2021-06-29T12:27:05Z) - UMIC: An Unreferenced Metric for Image Captioning via Contrastive
Learning [39.40274917797253]
In this paper, we introduce a new metric UMIC, an Unreferenced Metric for Image Captioning.
Based on Vision-and-Language BERT, we train UMIC to discriminate negative captions via contrastive learning.
Also, we observe critical problems of the previous benchmark dataset on image captioning metric, and introduce a new collection of human annotations on the generated captions.
arXiv Detail & Related papers (2021-06-26T13:27:14Z) - Intrinsic Image Captioning Evaluation [53.51379676690971]
We propose a learning based metrics for image captioning, which we call Intrinsic Image Captioning Evaluation(I2CE)
Experiment results show that our proposed method can keep robust performance and give more flexible scores to candidate captions when encountered with semantic similar expression or less aligned semantics.
arXiv Detail & Related papers (2020-12-14T08:36:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.