Javascript must be enabled to continue!
Benchmarking Multi-dimensional AIGC Video Quality Assessment: A Dataset and Unified Model
View through CrossRef
In recent years, AI-driven video generation has gained significant attention due to great advancements in visual and language generative techniques. Consequently, there is a growing need for accurate Video Quality Assessment (VQA) metrics to evaluate the perceptual quality of AI-generated content (AIGC) videos and optimize video generation models. However, assessing the quality of AIGC videos remains a significant challenge because these videos often exhibit highly complex distortions, such as unnatural actions and irrational objects. To address this challenge, we systematically investigate the AIGC-VQA problem in this article, considering both subjective and objective quality assessment perspectives. For the subjective perspective, we construct the
L
arge-scale
G
enerated
V
ideo
Q
uality Assessment (LGVQ) dataset, consisting of
\(2,\!808\)
AIGC videos generated by six video generation models using 468 carefully curated text prompts. Unlike previous subjective VQA experiments, we evaluate the perceptual quality of AIGC videos from three critical dimensions: spatial quality, temporal quality, and text-video alignment, which hold utmost importance for current video generation techniques. For the objective perspective, we establish a benchmark for evaluating existing quality assessment metrics on the LGVQ dataset. Our findings show that current metrics perform poorly on this dataset, highlighting a gap in effective evaluation tools. To bridge this gap, we propose the
U
nify
G
enerated
V
ideo
Q
uality Assessment (UGVQ) model, designed to accurately evaluate the multi-dimensional quality of AIGC videos. The UGVQ model integrates the visual and motion features of videos with the textual features of their corresponding prompts, forming a unified quality-aware feature representation tailored to AIGC videos. Experimental results demonstrate that UGVQ achieves state-of-the-art performance on the LGVQ dataset across all three quality dimensions, validating its effectiveness as an accurate quality metric for AIGC videos. We hope that our benchmark can promote the development of AIGC-VQA studies. Both the LGVQ dataset and the UGVQ model are publicly available on
https://github.com/zczhang-sjtu/UGVQ.git
.
Association for Computing Machinery (ACM)
Title: Benchmarking Multi-dimensional AIGC Video Quality Assessment: A Dataset and Unified Model
Description:
In recent years, AI-driven video generation has gained significant attention due to great advancements in visual and language generative techniques.
Consequently, there is a growing need for accurate Video Quality Assessment (VQA) metrics to evaluate the perceptual quality of AI-generated content (AIGC) videos and optimize video generation models.
However, assessing the quality of AIGC videos remains a significant challenge because these videos often exhibit highly complex distortions, such as unnatural actions and irrational objects.
To address this challenge, we systematically investigate the AIGC-VQA problem in this article, considering both subjective and objective quality assessment perspectives.
For the subjective perspective, we construct the
L
arge-scale
G
enerated
V
ideo
Q
uality Assessment (LGVQ) dataset, consisting of
\(2,\!808\)
AIGC videos generated by six video generation models using 468 carefully curated text prompts.
Unlike previous subjective VQA experiments, we evaluate the perceptual quality of AIGC videos from three critical dimensions: spatial quality, temporal quality, and text-video alignment, which hold utmost importance for current video generation techniques.
For the objective perspective, we establish a benchmark for evaluating existing quality assessment metrics on the LGVQ dataset.
Our findings show that current metrics perform poorly on this dataset, highlighting a gap in effective evaluation tools.
To bridge this gap, we propose the
U
nify
G
enerated
V
ideo
Q
uality Assessment (UGVQ) model, designed to accurately evaluate the multi-dimensional quality of AIGC videos.
The UGVQ model integrates the visual and motion features of videos with the textual features of their corresponding prompts, forming a unified quality-aware feature representation tailored to AIGC videos.
Experimental results demonstrate that UGVQ achieves state-of-the-art performance on the LGVQ dataset across all three quality dimensions, validating its effectiveness as an accurate quality metric for AIGC videos.
We hope that our benchmark can promote the development of AIGC-VQA studies.
Both the LGVQ dataset and the UGVQ model are publicly available on
https://github.
com/zczhang-sjtu/UGVQ.
git
.
Related Results
An optimisational model of benchmarking
An optimisational model of benchmarking
PurposeThe purpose of this paper is to develop a quantitative methodology for benchmarking process which is simple, effective and efficient as a rejoinder to benchmarking detractor...
A review on benchmarking of supply chain performance measures
A review on benchmarking of supply chain performance measures
PurposeThe purpose of this paper is to redress the imbalances in the past literature of supply chain benchmarking and enhance data envelopment analysis (DEA) modeling approach in s...
The Advanced Path of Research on Digital Display Empowered by AIGC Technology for Intangible Cultural Heritage
The Advanced Path of Research on Digital Display Empowered by AIGC Technology for Intangible Cultural Heritage
Artificial Intelligence Generated Content (AIGC) technology combines machine learning and natural language processing to provide a new way to digitally display and disseminate inta...
Enhancing Real-Time Video Processing With Artificial Intelligence: Overcoming Resolution Loss, Motion Artifacts, And Temporal Inconsistencies
Enhancing Real-Time Video Processing With Artificial Intelligence: Overcoming Resolution Loss, Motion Artifacts, And Temporal Inconsistencies
Purpose: Traditional video processing techniques often struggle with critical challenges such as low resolution, motion artifacts, and temporal inconsistencies, especially in real-...
The need for adaptive processes of benchmarking in small business‐to‐business services
The need for adaptive processes of benchmarking in small business‐to‐business services
PurposeThis paper aims to explore current management attitudes towards benchmarking and its implementation within small business‐to‐business service firms in order to enhance a dee...
Factors Influencing Preschool Teachers' Continuous Intention to Use AIGC in Education
Factors Influencing Preschool Teachers' Continuous Intention to Use AIGC in Education
Artificial Intelligence Generated Content (AIGC) is becoming a valuable tool ineducation. It supports personalized, interactive, and scalable learning. However, little is known abo...
AIGC Technology Adoption in Yixing Zisha Ceramic Design: An Extended TAM Model Analysis
AIGC Technology Adoption in Yixing Zisha Ceramic Design: An Extended TAM Model Analysis
This study examines the willingness of the Yixing Zisha design industry to adopt Artificial Intelligence Generated Content (AIGC) technology, using the extended Technology Acceptan...
Exploring Designer Trust in Artificial Intelligence-Generated Content: TAM/TPB Model Study
Exploring Designer Trust in Artificial Intelligence-Generated Content: TAM/TPB Model Study
Traditionally, users have perceived that only manual laborers or those in repetitive jobs would be subject to technological substitution. However, with the emergence of technologie...

