- 2026.02.06 The code and pretrained model of RALI and VQ-Insight are released!
- 2026.02.06 RALI has been accepted at ICLR 2026 as an oral presentation!
- 2025.11.08 VQ-Insight has been accepted at AAAI 2026 as an oral presentation!
- 2025.09.19 Q-Insight has been accepted at NeurIPS 2025 as a spotlight (Top 3%)!
- 2025.05.30 Released training and testing code, along with the pretrained model.
- 2025.05.26 Released our v2 paper.
- 2025.03.28 Released the Q-Insight technical report.
Weiqi Li, Xuanyu Zhang, Shijie Zhao, Yabin Zhang, Junlin Li, Li Zhang and Jian Zhang
PLCC comparisons between our proposed Q-Insight and existing IQA metrics (left) and three example applications of our Q-Insight (right) are presented. Q-Insight demonstrates significantly improved performance compared to existing methods, especially on out-of-domain datasets. Additionally, Q-Insight effectively supports quality score regression, image degradation perception, and zero-shot image comparison reasoning tasks.
VQ-Insight: Teaching VLMs for AI-Generated Video Quality Understanding via Progressive Visual Reinforcement Learning
Xuanyu Zhang*, Weiqi Li*, Shijie Zhao, Junlin Li, Li Zhang, Jian Zhang
We propose a reasoning-style vision-language model VQ-Insight, which accurately performs AIGC video preference comparison, AIGC video multi-dimension scoring, and natural video scoring, accompanied by detailed and reasonable reasoning processes. Our VQ-Insight can be applied to post-training of video generation models and zero-shot content repairing.
Shijie Zhao*, Xuanyu Zhang*, Weiqi Li, Junlin Li, Li Zhang, Tianfan Xue, Jian Zhang
We revisit the reasoning mechanism in MLLM-based IQA model (such as Q-Insight) and propose a CLIP-based lightweight image scorer RALI. We verifies that through RL training, MLLMs leverage their reasoning capability to convert redundant visual representations into compact, cross-domain aligned text representations. This conversion is the source of the generalization exhibited by these reasoning-based IQA models. RALI uses only about 4% of Q-Insightβs parameters and inference time, while achieving comparable accuracy.
git clone https://github.com/bytedance/Q-Insight.git
bash setup.shTo run VQ-Insight, install additional pacakages.
cd src/eval/qwen-vl-utils
pip install -e .[decord]cd src/eval/
python demo_score.py
cd src/eval/
python demo_dist.py
cd src/eval/
python demo_comparison.py
cd src/eval/
python demo_vqinsight_score.py \
--video_path "../../assets/demo_natural.mp4" \
--video_type naturalcd src/eval/
python demo_vqinsight_score.py \
--video_path "../../assets/demo_aigc.mp4" \
--video_type aigccd src/eval/
python demo_vqinsight_comp.py \
--video_a "../../assets/demo_comp1.mp4" \
--video_b "../../assets/demo_comp2.mp4" \
--model_name_or_path Bytedance/Q-Insight
Please download the RALI pretrained weights from the link. After downloading, place the checkpoint under Q-Insight/checkpoints, so that the directory structure becomes:
Q-Insight/
βββ checkpoints/
β βββ ckpt.pt
β βββ pca.pkl
β βββ basis.npz
β βββ best/
β βββ config.json
β βββ pytorch_model.bin (or *.safetensors)
β βββ preprocessor_config.json
β βββ ...
βββ src/
βββ assets/
βββ README.md
Then run the following code:
cd src/eval/
python demo_rali_score.pyDownload meta files from Data-DeQA-Score and the source images from the KONIQ dataset.
Arrange the folders in ./src/open-r1-multimodal/dataas follows:
|-- Data-DeQA-Score
|-- KONIQ
|-- images/*.jpg
|-- metas
Download the refA_sd_brief subset from KADIS-700K.
Arrange the folders in ./src/open-r1-multimodal/data as follows:
|-- KADIS-700K
|-- refA_sd_brief
|-- dist_imgs/*.jpg
|-- metas/train_dist.json
Download the validation dataset of DiffIQA.
Arrange the folders in ./src/open-r1-multimodal/data as follows:
|-- DiffIQA
|-- ValidationImage
|-- images
|-- train_comparison.json
cd src/open-r1-multimodal/
bash run_qinsight_score_and_dist.sh
cd src/open-r1-multimodal/
bash run_qinsight_comparison.sh
- Release the code and model of VQ-Insight
- Add support for LoRA fine-tuning
- Provide a Gradio demo
- Release inference code and weights
- Release training code
- Release the paper
We appreciate the releasing codes and data of VLM-R1, DepictQA and DeQA-Score.
If Q-Insight Family is helpful, please help to β the repo.
If you find the code helpful in your research or work, please cite the following papers:
@article{li2025qinsight,
title={Q-Insight: Understanding Image Quality via Visual Reinforcement Learning},
author={Li, Weiqi and Zhang, Xuanyu and Zhao, Shijie and Zhang, Yabin and Li, Junlin and Zhang, Li and Zhang, Jian},
journal={Proceedings of the Advances in Neural Information Processing Systems (NeurIPS)},
year={2025}
}
@article{zhang2025vqinsight,
title={VQ-Insight: Teaching VLMs for AI-Generated Video Quality Understanding via Progressive Visual Reinforcement Learning},
author={Zhang, Xuanyu and Li, Weiqi and Zhao, Shijie and Li, Junlin and Zhang, Li and Zhang, Jian},
journal={Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)},
year={2026}
}
@article{zhao2025reasoning,
title={Reasoning as Representation: Rethinking Visual Reinforcement Learning in Image Quality Assessment},
author={Zhao, Shijie and Zhang, Xuanyu and Li, Weiqi and Li, Junlin and Zhang, Li and Xue, Tianfan and Zhang, Jian},
journal={Proceedings of the International Conference on Learning Representations (ICLR)},
year={2026}
}


