Evaluating Automatically Generated Phoneme Captions for Images
- URL: http://arxiv.org/abs/2007.15916v1
- Date: Fri, 31 Jul 2020 09:21:13 GMT
- Title: Evaluating Automatically Generated Phoneme Captions for Images
- Authors: Justin van der Hout, Zolt\'an D'Haese, Mark Hasegawa-Johnson, Odette
Scharenborg
- Abstract summary: Image2Speech is the relatively new task of generating a spoken description of an image.
This paper presents an investigation into the evaluation of this task.
BLEU4 is the best currently existing metric for the Image2Speech task.
- Score: 44.20957732654963
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image2Speech is the relatively new task of generating a spoken description of
an image. This paper presents an investigation into the evaluation of this
task. For this, first an Image2Speech system was implemented which generates
image captions consisting of phoneme sequences. This system outperformed the
original Image2Speech system on the Flickr8k corpus. Subsequently, these
phoneme captions were converted into sentences of words. The captions were
rated by human evaluators for their goodness of describing the image. Finally,
several objective metric scores of the results were correlated with these human
ratings. Although BLEU4 does not perfectly correlate with human ratings, it
obtained the highest correlation among the investigated metrics, and is the
best currently existing metric for the Image2Speech task. Current metrics are
limited by the fact that they assume their input to be words. A more
appropriate metric for the Image2Speech task should assume its input to be
parts of words, i.e. phonemes, instead.
Related papers
- InfoMetIC: An Informative Metric for Reference-free Image Caption
Evaluation [69.1642316502563]
We propose an Informative Metric for Reference-free Image Caption evaluation (InfoMetIC)
Given an image and a caption, InfoMetIC is able to report incorrect words and unmentioned image regions at fine-grained level.
We also construct a token-level evaluation dataset and demonstrate the effectiveness of InfoMetIC in fine-grained evaluation.
arXiv Detail & Related papers (2023-05-10T09:22:44Z) - Zero-Shot Video Captioning with Evolving Pseudo-Tokens [79.16706829968673]
We introduce a zero-shot video captioning method that employs two frozen networks: the GPT-2 language model and the CLIP image-text matching model.
The matching score is used to steer the language model toward generating a sentence that has a high average matching score to a subset of the video frames.
Our experiments show that the generated captions are coherent and display a broad range of real-world knowledge.
arXiv Detail & Related papers (2022-07-22T14:19:31Z) - Are metrics measuring what they should? An evaluation of image
captioning task metrics [0.21301560294088315]
Image Captioning is a current research task to describe the image content using the objects and their relationships in the scene.
To tackle this task, two important research areas are used, artificial vision, and natural language processing.
We present an evaluation of several kinds of Image Captioning metrics and a comparison between them using the well-known MS COCO dataset.
arXiv Detail & Related papers (2022-07-04T21:51:47Z) - Transparent Human Evaluation for Image Captioning [70.03979566548823]
We develop a rubric-based human evaluation protocol for image captioning models.
We show that human-generated captions show substantially higher quality than machine-generated ones.
We hope that this work will promote a more transparent evaluation protocol for image captioning.
arXiv Detail & Related papers (2021-11-17T07:09:59Z) - Can Audio Captions Be Evaluated with Image Caption Metrics? [11.45508807551818]
We propose a metric named FENSE, where we combine the strength of Sentence-BERT in capturing similarity, and a novel Error Detector to penalize erroneous sentences for robustness.
On the newly established benchmarks, FENSE outperforms current metrics by 14-25% accuracy.
arXiv Detail & Related papers (2021-10-10T02:34:40Z) - Contrastive Semantic Similarity Learning for Image Captioning Evaluation
with Intrinsic Auto-encoder [52.42057181754076]
Motivated by the auto-encoder mechanism and contrastive representation learning advances, we propose a learning-based metric for image captioning.
We develop three progressive model structures to learn the sentence level representations.
Experiment results show that our proposed method can align well with the scores generated from other contemporary metrics.
arXiv Detail & Related papers (2021-06-29T12:27:05Z) - Intrinsic Image Captioning Evaluation [53.51379676690971]
We propose a learning based metrics for image captioning, which we call Intrinsic Image Captioning Evaluation(I2CE)
Experiment results show that our proposed method can keep robust performance and give more flexible scores to candidate captions when encountered with semantic similar expression or less aligned semantics.
arXiv Detail & Related papers (2020-12-14T08:36:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.