- 2024-03-03: Upload an example for KVQ competition based on SimpleVQA in this GitHub.
- 2024-03-26: Upload the validation metric for competition named "metric_score.py" and release the quality label for the validation dataset in link.
- 2024-03-26: We open the submission for the testing dataset in Codalab, you can continue to validate your methods by submitting the results.
- 2024-07-27: The KSVQE of our work has been released!!
Short-form UGC videos, like Kwai and TikTok, have become a popular and irreplaceable mainstream media form, thriving on user-friendly engagement, and kaleidoscope creation, etc. However, the advancing content generation modes, e.g., special effects, and sophisticated processing workflows, e.g., de-artifacts, have introduced significant challenges to recent UGC video quality assessment (VQA).
To promote the development of short-form UGC VQA, we have the first NTIRE challenge on Short-form UGC Video Quality Assessment, jointly with the NTIRE workshop. This competition introduces the large-scale Kaleidoscope short Video database for Quality assessment, termed KVQ, for training and evaluation. The KVQ database compromises 600 user-uploaded short videos and 3600 processed videos through the diverse practical processing workflows. Moreover, it contains nine primary content scenarios in the practical short-form video platform, including landscape, crowd, person, food, portrait, computer graphic (termed as CG), caption, and stage, covering almost all existing creation modes and scenarios, and the ratio of each category of content satisfies the practical online statistics. The quality score of each short-form video and the partial ranked score are annotated with professional researchers on image processing.
The goal of this competition is to establish a new and applicable benchmark for short-form User Generated Content (UGC) quality assessment. We are looking forward to the collaborative efforts of our participants, aiming to elevate the quality of short-form video content. This competition will introduce an innovative evaluation methodology that incorporates a coarse-grained quality score measurement, along with fine-grained rankings for more challenging samples.
conda create -n KVQ python=3.8
conda activate KVQ
pip install -r requirements.txt
Our codes are compatible with pytorch1.9.0, you may try newer version.
Download KVQ dataset from codalab competition [this link] (https://codalab.lisn.upsaclay.fr/) Please add the path of KVQ and annotation to the items of "data_prefix" and "anno_file" in the config file (i.e. /config/kwai_simpleVQA.yml)
python SlowFast_features.py --gpu_ids 0,1 --video_root yout_path --video_csv yout_path
Please add the path of Slowfast feature to the items of "data_prefix_3D" in the config file (i.e. /config/kwai_simpleVQA.yml)
nohup python -u train.py --o config/kwai_simpleVQA.yml --gpu_id 0,1 > log/kwai_simpleVQA.log 2>&1 &
or
bash scripts/train.sh
nohup python -u test.py --o config/kwai_simpleVQA_test.yml --gpu_id 0 > log/kwai_simpleVQA_test.log 2>&1 &
or
bash scripts/test.sh
CUDA_VISIBLE_DEVICES=0,1,2,3 nohup python -m torch.distributed.launch --nproc_per_node=4 --master_port 3332 --use_env train_ddp.py --o config/Kwai_KSVQE.yml --gpu_id 0,1,2,3 -r checkpoint_ddp/ > log/Kwai_KSVQE_ddp_loadpretrained.log 2>&1 &
bash scripts/train_KSVQE_ddp.sh
nohup python -u train.py --o config/Kwai_KSVQE.yml -r checkpoint_dp/ --gpu_id 0,1 > log/Kwai_KSVQE.log 2>&1 &
or bash scripts/train_KSVQE.sh
python /data2/luyt/KSVQE/test.py --o config/Kwai_KSVQE_test.yml --gpu_id 7```
or
```bash
bash scripts/test_KSVQE.sh
Please download the pertrained weight of 3D swin transformer backbone on LSVQ in this link: https://drive.google.com/file/d/1HcbqCRRnVLnh738VHkpz7UMNuM3v59QU/view?usp=drive_link
Please cite us if this work is helpful to you.
@inproceedings{lu2024kvq,
title={KVQ: Kwai Video Quality Assessment for Short-form Videos},
author={Lu, Yiting and Li, Xin and Pei, Yajing and Yuan, Kun and Xie, Qizhi and Qu, Yunpeng and Sun, Ming and Zhou, Chao and Chen, Zhibo},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
year={2024}
}
@inproceedings{li2024ntire,
title={NTIRE 2024 Challenge on Short-form UGC Video Quality Assessment: Methods and Results},
author={Li, Xin and Yuan, Kun and Pei, Yajing and Lu, Yiting and Sun, Ming and Zhou, Chao and Chen, Zhibo and Timofte, Radu and others},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year={2024}
}
The basic code is partially from the below repos.