Blind or no-reference video quality assessment of user-generated content (UGC) has become a trending, challenging, heretofore unsolved problem. Accurate and efficient video quality predictors suitable for this content are thus in great demand to achieve more intelligent analysis and processing of UGC videos. Previous studies have shown that natural scene statistics and deep learning features are both sufficient to capture spatial distortions, which contribute to a significant aspect of UGC video quality issues. However, these models are either incapable or inefficient for predicting the quality of complex and diverse UGC videos in practical applications. Here we introduce an effective and efficient video quality model for UGC content, which we dub the Rapid and Accurate Video Quality Evaluator (RAPIQUE), which we show performs comparably to state-of-the-art (SOTA) models but with orders-of-magnitude faster runtime. RAPIQUE combines and leverages the advantages of both quality-aware scene statistics features and semantics-aware deep convolutional features, allowing us to design the first general and efficient spatial and temporal (space-time) bandpass statistics model for video quality modeling. Our experimental results on recent large-scale UGC video quality databases show that RAPIQUE delivers top performances on all the datasets at a considerably lower computational expense. We hope this work promotes and inspires further efforts towards practical modeling of video quality problems for potential real-time and low-latency applications. To promote public usage, an implementation of RAPIQUE has been made freely available online: \url{https://github.com/vztu/RAPIQUE}.
翻译:对用户生成的内容(UGC)进行盲目或无参考视频质量评估已成为一个趋势、挑战性、迄今尚未解决的问题。因此,非常需要准确而高效的适合该内容的视频质量预测器,以便更明智地分析和处理UGC视频。先前的研究显示,自然现场统计和深层次学习特征都足以捕捉空间扭曲,从而导致UGC视频质量问题的一个重要方面。然而,这些模型在实际应用中无法预测复杂和多样的UGC视频质量,或者缺乏效率。在这里,我们为UGC内容引入一个高效益和高效率的视频质量模型,我们调用快速和准确的视频质量预测器,我们用它来进行更智能的分析和处理。