Abstract
In recent years, image generation technology has rapidly advanced, resulting in the creation of a vast array of AI-generated images (AIGIs). However, the quality of these AIGIs is highly inconsistent, with low-quality AIGIs severely impairing the visual experience of users. Due to the widespread application of AIGIs, the AI-generated image quality assessment (AIGIQA), aimed at evaluating the quality of AIGIs from the perspective of human perception, has garnered increasing interest among scholars. Nonetheless, current research has not yet fully explored this field. We have observed that existing databases are limited to images generated from single scenario settings. Databases such as AGIQA-1K, AGIQA-3K, and AIGCIQA2023, for example, only include images generated by text-to-image generative models. This oversight highlights a critical gap in the current research landscape, underscoring the need for dedicated databases catering to image-to-image scenarios, as well as more comprehensive databases that encompass a broader range of AI-generated image scenarios. Addressing these issues, we have established a large scale perceptual quality assessment database for both text-to-image and image-to-image AIGIs, named PKU-AIGIQA-4K. We then conduct a well-organized subjective experiment to collect quality labels for AIGIs and perform a comprehensive analysis of the PKU-AIGIQA-4K database. Regarding the use of image prompts during the training process, we propose three image quality assessment (IQA) methods based on pre-trained models that include a no-reference method NR-AIGCIQA, a full-reference method FR-AIGCIQA, and a partial-reference method PR-AIGCIQA. Finally, leveraging the PKU-AIGIQA-4K database, we conduct extensive benchmark experiments and compare the performance of the proposed methods and the current IQA methods.
Abstract (translated)
近年来,图像生成技术快速发展,导致产生了大量AI生成的图像(AIGIs)。然而,这些AIGIs的质量差异很大,低质量的AIGIs严重地损害了用户的使用体验。由于AIGIs的广泛应用,从人类的角度评估AIGI质量的人工智能图像质量评估(AIGIQA)受到了越来越多的关注。然而,目前的 research 尚未完全探索这个领域。我们观察到,现有的数据库仅限于从单一场景设置生成的图像。例如,AGIQA-1K,AGIQA-3K和AIGCIQA2023等数据库仅包括由文本到图像生成模型的图像。这一缺陷突显了当前研究格局中的关键空白,强调了需要针对图像到图像场景建立专门的数据库以及更全面的涵盖更广泛AI生图像场景的数据库。为解决这些问题,我们建立了一个大规模的主观质量评估数据库,名为PKU-AIGIQA-4K。然后,我们进行了一个组织良好的主观实验,收集了AIGIs的质量标签,并对PKU-AIGIQA-4K数据库进行了全面分析。关于在训练过程中使用图像提示的问题,我们提出了三种基于预训练模型的图像质量评估(IQA)方法,包括无参考方法NR-AIGCIQA,完整参考方法FR-AIGCIQA和部分参考方法PR-AIGCIQA。最后,利用PKU-AIGIQA-4K数据库,我们进行了广泛的基准实验,并比较了所提出方法和现有IQA方法的性能。
URL
https://arxiv.org/abs/2404.18409