Zheng Chen, Xun Zhang, Wenbo Li, Renjing Pei, Fenglong Song, Xiongkuo Min, Xiaohong Liu, Xin Yuan, Yong Guo, and Yulun Zhang, "Grounding-IQA: Multimodal Language Grounding Model for Image Quality Assessment", 2024
[project] [arXiv] [supplementary material] [dataset] [pretrained models]
- 2024-11-26: This repo is released.
Abstract: The development of multimodal large language models (MLLMs) enables the evaluation of image quality through natural language descriptions. This advancement allows for more detailed assessments. However, these MLLM-based IQA methods primarily rely on general contextual descriptions, sometimes limiting fine-grained quality assessment. To address this limitation, we introduce a new image quality assessment (IQA) task paradigm, grounding-IQA. This paradigm integrates multimodal referring and grounding with IQA to realize more fine-grained quality perception. Specifically, grounding-IQA comprises two subtasks: grounding-IQA-description (GIQA-DES) and visual question answering (GIQA-VQA). GIQA-DES involves detailed descriptions with precise locations (e.g., bounding boxes), while GIQA-VQA focuses on quality QA for local regions. To realize grounding-IQA, we construct a corresponding dataset, GIQA-160K, through our proposed automated annotation pipeline. Furthermore, we develop a well-designed benchmark, GIQA-Bench. The benchmark comprehensively evaluates the model grounding-IQA performance from three perspectives: description quality, VQA accuracy, and grounding precision. Experiments demonstrate that our proposed task paradigm, dataset, and benchmark facilitate the more fine-grained IQA application.
- Release testing and training code.
- Release GIQA-160K and GIQA-Bench.
- Release pre-trained models.
- Provide WebUI.
- Provide HuggingFace demo.
- Datasets
- Models
- Training
- Testing
- Results
- Citation
- Acknowledgements
We achieve impressive performance on GIQA-DES and GIQA-VQA tasks.
Qualitative Results (click to expand)
- Results in Fig. 7 of the main paper
If you find the code helpful in your research or work, please cite the following paper(s).
@article{chen2024grounding,
title={Grounding-IQA: Multimodal Language Grounding Model for Image Quality Assessment},
author={Chen, Zheng and Zhang, Xun and Li, Wenbo and Pei, Renjing and Song, Fenglong and Min, Xiongkuo and Liu, Xiaohong and Yuan, Xin and Guo, Yong and Zhang, Yulun},
journal={arXiv preprint arXiv:2411.17237},
year={2024}
}
This project is based on Q-Instruct, DepictQA, mPLUG-Owl, and LLaVA.