Evaluating Text-to-Image Synthesis: Survey and Taxonomy of Image Quality Metrics

Sebastian Hartwig

Ulm University

Dominik Engel

Ulm University

Leon Sick

Ulm University

Hannah Kniesel

Ulm University

Tristan Payer

Ulm University

Poonam Poonam

Ulm University

Michael Glöckler

Ulm University

Alex Bäuerle

Carnegie Mellon University

Timo Ropinski

Ulm University

https://arxiv.org/abs/2403.11821 2024

Abstract

Recent advances in text-to-image synthesis enabled through a combination of language and vision foundation models have led to a proliferation of the tools available and an increased attention to the field. When conducting text-to-image synthesis, a central goal is to ensure that the content between text and image is aligned. As such, there exist numerous evaluation metrics that aim to mimic human judgement. However, it is often unclear which metric to use for evaluating text-to-image synthesis systems as their evaluation is highly nuanced. In this work, we provide a comprehensive overview of existing text-to-image evaluation metrics. Based on our findings, we propose a new taxonomy for categorizing these metrics. Our taxonomy is grounded in the assumption that there are two main quality criteria, namely compositionality and generality, which ideally map to human preferences. Ultimately, we derive guidelines for practitioners conducting text-to-image evaluation, discuss open challenges of evaluation mechanisms, and surface limitations of current metrics.

Bibtex

@preprint{hartwig2024survey,
	title={Evaluating Text-to-Image Synthesis: Survey and Taxonomy of Image Quality Metrics},
	author={Hartwig, Sebastian and Engel, Dominik and Sick, Leon and Kniesel, Hannah and Payer, Tristan and Poonam, Poonam and Gl{\"o}ckler, Michael and B{\"a}uerle, Alex and Ropinski, Timo},
	year={2024},
	journal={arxiv preprint arXiv:https://arxiv.org/abs/2403.11821}
}