Revisiting MLLM Based Image Quality Assessment: Errors and Remedy
DOI:
https://doi.org/10.1609/aaai.v40i11.37908Abstract
The rapid progress of multi-modal large language models (MLLMs) has boosted the task of image quality assessment (IQA). However, a key challenge arises from the inherent mismatch between the discrete token outputs of MLLMs and the continuous nature of quality scores required by IQA tasks. This discrepancy significantly hinders the performance of MLLM-based IQA methods. Previous approaches that convert discrete token predictions into continuous scores often suffer from conversion errors. Moreover, the semantic confusion introduced by level tokens (e.g., “good”) further constrains the performance of MLLMs on IQA tasks and degrades their original capabilities to related tasks. To tackle these problems, we provide a theoretical analysis of the errors inherent in previous approaches and, motivated by this analysis, propose a simple yet effective framework, Q-Scorer. This framework incorporates a lightweight regression module and IQA-specific score tokens into the MLLM pipeline. Extensive experiments demonstrate that Q-Scorer achieves state-of-the-art performance across multiple IQA benchmarks, generalizes well to mixed datasets, and further improves combined with other methods.Published
2026-03-14
How to Cite
Tang, Z., Yang, S., Peng, B., Wang, Z., & Dong, J. (2026). Revisiting MLLM Based Image Quality Assessment: Errors and Remedy. Proceedings of the AAAI Conference on Artificial Intelligence, 40(11), 9475–9483. https://doi.org/10.1609/aaai.v40i11.37908
Issue
Section
AAAI Technical Track on Computer Vision VIII