Image2Text2Image: A Novel Framework for Label-Free Evaluation of Image-to-Text Generation with Text-to-Image Diffusion Models

Open Access
Authors
Publication date 2025
Host editors
  • I. Ide
  • I. Kompatsiaris
  • C. Xu
  • K. Yanai
  • W.-T. Chu
  • N. Nitta
  • M. Riegler
  • T. Yamasaki
Book title MultiMedia Modeling
Book subtitle 31st International Conference on Multimedia Modeling, MMM 2025, Nara, Japan, January 8–10, 2025 : proceedings
ISBN
  • 9789819620708
ISBN (electronic)
  • 9789819620715
Series Lecture Notes in Computer Science
Event 31st International Conference on Multimedia Modeling, MMM 2025
Volume | Issue number IV
Pages (from-to) 413-427
Publisher Singapore: Springer
Organisations
  • Faculty of Economics and Business (FEB) - Amsterdam Business School Research Institute (ABS-RI)
  • Faculty of Law (FdR)
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Evaluating the quality of automatically generated image descriptions is a complex task that requires metrics capturing various dimensions, such as grammaticality, coverage, accuracy, and truthfulness. Although human evaluation provides valuable insights, its cost and time-consuming nature pose limitations. Existing automated metrics like BLEU, ROUGE, METEOR, and CIDEr attempt to fill this gap, but they often exhibit weak correlations with human judgment. To address this challenge, we propose a novel evaluation framework called Image2Text2Image, which leverages diffusion models, such as Stable Diffusion or DALL-E, for text-to-image generation. In the Image2Text2Image framework, an input image is first processed by a selected image captioning model, chosen for evaluation, to generate a textual description. Using this generated description, a diffusion model then creates a new image. By comparing features extracted from the original and generated images, we measure their similarity using a designated similarity metric. A high similarity score suggests that the model has produced a faithful textual description, while a low score highlights discrepancies, revealing potential weaknesses in the model’s performance. Notably, our framework does not rely on human-annotated reference captions, making it a valuable tool for assessing image captioning models. Extensive experiments and human evaluations validate the efficacy of our proposed Image2Text2Image evaluation framework.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-981-96-2071-5_30
Downloads
Image2Text2Image (Final published version)
Permalink to this page
Back