arxiv_dump / txt /2204.14047.txt
billxbf's picture
Upload 101 files
8f1929a verified
raw
history blame
58.6 kB
A Deep Learning based No-reference Quality Assessment Model
for UGC Videos
Wei Sun
Shanghai Jiao Tong University
Shanghai, China
[email protected] Min
Shanghai Jiao Tong University
Shanghai, China
[email protected]
Wei Lu
Shanghai Jiao Tong University
Shanghai, China
[email protected] Zhai∗
Shanghai Jiao Tong University
Shanghai, China
[email protected]
ABSTRACT
Quality assessment for User Generated Content (UGC) videos plays
an important role in ensuring the viewing experience of end-users.
Previous UGC video quality assessment (VQA) studies either use
the image recognition model or the image quality assessment (IQA)
models to extract frame-level features of UGC videos for quality
regression, which are regarded as the sub-optimal solutions be-
cause of the domain shifts between these tasks and the UGC VQA
task. In this paper, we propose a very simple but effective UGC
VQA model, which tries to address this problem by training an
end-to-end spatial feature extraction network to directly learn the
quality-aware spatial feature representation from raw pixels of the
video frames. We also extract the motion features to measure the
temporal-related distortions that the spatial features cannot model.
The proposed model utilizes very sparse frames to extract spatial
features and dense frames (i.e. the video chunk) with a very low
spatial resolution to extract motion features, which thereby has
low computational complexity. With the better quality-aware fea-
tures, we only use the simple multilayer perception layer (MLP)
network to regress them into the chunk-level quality scores, and
then the temporal average pooling strategy is adopted to obtain
the video-level quality score. We further introduce a multi-scale
quality fusion strategy to solve the problem of VQA across differ-
ent spatial resolutions, where the multi-scale weights are obtained
from the contrast sensitivity function of the human visual system.
The experimental results show that the proposed model achieves
the best performance on five popular UGC VQA databases, which
demonstrates the effectiveness of the proposed model. The code is
available at https://github.com/sunwei925/SimpleVQA.
CCS CONCEPTS
•Computing methodologies →Modeling methodologies .
∗Corresponding author: Guangtao Zhai.
Permission to make digital or hard copies of all or part of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed
for profit or commercial advantage and that copies bear this notice and the full citation
on the first page. Copyrights for components of this work owned by others than ACM
must be honored. Abstracting with credit is permitted. To copy otherwise, or republish,
to post on servers or to redistribute to lists, requires prior specific permission and/or a
fee. Request permissions from [email protected].
MM ’22, October 10–14, 2022, Lisboa, Portugal
©2022 Association for Computing Machinery.
ACM ISBN 978-1-4503-9203-7/22/10. . . $15.00
https://doi.org/10.1145/3503161.3548329KEYWORDS
video quality assessment, UGC videos, deep learning, feature fusion
ACM Reference Format:
Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao Zhai∗. 2022. A Deep Learn-
ing based No-reference Quality Assessment Model for UGC Videos. In Pro-
ceedings of the 30th ACM International Conference on Multimedia (MM ’22),
October 10–14, 2022, Lisboa, Portugal. ACM, New York, NY, USA, 10 pages.
https://doi.org/10.1145/3503161.3548329
1 INTRODUCTION
With the proliferation of mobile devices and wireless networks in
recent years, User Generated Content (UGC) videos have exploded
over the Internet. It has become a popular daily activity for the gen-
eral public to create, view, and share UGC videos through various
social media applications such as YouTube, TikTok, etc. However,
UGC videos are captured by a wide variety of consumers, ranging
from professional photographers to amateur users, which makes
the visual quality of UGC videos vary greatly. In order to ensure
the Quality of Experience (QoE) of end-users, the service providers
need to monitor the quality of UGC videos in the entire streaming
media link, including but not limited to video uploading, compress-
ing, post-processing, transmitting, etc. Therefore, with billions of
video viewing and millions of newly uploaded UGC videos every
day, an effective and efficient video quality assessment (VQA) model
is needed to measure the perceptual quality of UGC videos.
Objective VQA can be divided into full-reference (FR), reduced-
reference (RR), and no-reference (NR) according to the amount of
pristine video information needed. Since there is no reference video
for in-the-wild UGC videos, only NR VQA models are qualified for
evaluating their quality. Although NR VQA algorithms [ 21,23,26]
have been studied for many years, most of them were developed
for Professionally Generated Content (PGC) videos with synthetic
distortions, where the pristine PGC videos are shot by photogra-
phers using professional devices and are normally of high quality,
and the distorted PGC videos are then degraded by specific video
processing algorithms such as video compression, transmission, etc.
So, previous VQA studies mainly focus on modeling several types
of distortions caused by specific algorithms, which makes them less
effective for UGC videos with in-the-wild distortions. To be more
specific, the emerging UGC videos pose the following challenges
to the existing VQA algorithms for PGC videos:arXiv:2204.14047v2 [cs.CV] 20 Oct 2022MM ’22, October 10–14, 2022, Lisboa, Portugal Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao Zhai∗
First, the distortion types of UGC videos are diverse. A mass
of UGC videos are captured by amateur users, which may suffer
various distortion types such as under/over exposure, low visibility,
jitter, noise, color shift, etc. These authentic distortions are intro-
duced in the shooting processing and cannot be modeled by the
single distortion type, which thereby requires that the VQA models
have a more strong feature representation ability to qualify the
authentic distortions. Second, the content and forms of UGC videos
are extremely rich. UGC videos can be natural scenes, animation
[35], games [ 46,47], screen content, etc. Note that the statistics
characteristics of different video content vary greatly. For example,
the natural scenes statistics (NSS) features [ 22–24,26] are com-
monly used in the previous VQA studies to measure the distortions
of natural scene content, but they may be ineffective for computer-
generated content like animation or games. In addition, live videos,
videoconferencing, etc. are also ubiquitous for UGC videos nowa-
days, whose quality is severely affected by the network bandwidth.
Third, due to the advancement of shooting devices, more high res-
olution [ 18] and high frame rate [ 19,52,53] videos have emerged
on the Internet. The various kinds of resolutions and frame rates
are also important factors for video quality. What’s more, users
can view the UGC videos through mobile devices anywhere and at
any time, so the display [ 25] and the viewing environment such as
ambient luminance [ 29], etc. also affect the perceptual quality of
UGC videos to a certain extent. However, these factors are rarely
considered by previous studies.
The recently released large-scale UGC VQA databases such as
KoNViD-1k [ 8], YouTube UGC [ 36], LSVQ [ 44], etc. have greatly
promoted the development of UGC VQA. Several deep learning
based NR VQA models [ 14,15,37,40,44] have been proposed to
solve some challenges mentioned above and achieve pretty good
performance. However, there are still some problems that need to
be addressed. First, the previous studies either use the image recog-
nition model [ 15][44] or the pretrained image quality assessment
(IQA) models [ 37][40][14] to extract frame-level features, which
lacks an end-to-end learning method to learn the quality-aware
spatial feature representation from raw pixels of video frames. Sec-
ond, previous studies usually extract the features from all video
frames and have a very high computational complexity, making
them difficult to apply to real-world scenarios. Since there is much
redundancy spatial information between the adjacent frames, we
argue that there is not necessary to extract the features from all
frames. Third, the spatial resolution and frame rate of UGC videos
as well as other factors such as the display, viewing environment,
etc. are still rarely considered by these studies. However, these
factors are very important for the perceptual quality of UGC videos
since the contrast sensitivity of the human visual system (HVS) is
affected by them.
In this paper, to address the challenges mentioned above, we
propose a very simple but effective deep learning based VQA model
for UGC videos. The proposed framework is illustrated in Figure 1,
which consists of the feature extraction module, the quality regres-
sion module, and the quality pooling module. For the feature ex-
traction module, we extract quality-aware features from the spatial
domain and the spatial-temporal domain to respectively measure
the spatial distortions and motion distortions. Instead of using thepretrained model to extract the spatial features in the previous stud-
ies, we propose to train an end-to-end spatial feature extraction
network to learn quality-aware feature representation in the spatial
domain, which thereby makes full use of various video content
and distortion types in current UGC VQA databases. We then uti-
lize the action recognition network to extract the motion features,
which can make up the temporal-related distortions that the spatial
features cannot model. Considering that the spatial features are
sensitive to the resolution while the motion features are sensitive
to the frame rate, we first split the video into continuous chunks
and then extract the spatial features and motion features by using a
key frame of each chunk and all frames of each chunk but at a low
spatial resolution respectively. So, the computational complexity of
the proposed model can be greatly reduced.
For the quality regression module, we use the multilayer per-
ception (MLP) network to map the quality-aware features into
the chunk-level quality scores, and the temporal average pooling
strategy is adopted to obtain the final video quality. In order to
solve the problem of quality assessment across different resolu-
tions, we introduce a multi-scale quality fusion strategy to fuse
the quality scores of the videos with different resolutions, where
the multi-scale weights are obtained from the contrast sensitivity
function (CSF) of HVS by considering the viewing environment
information. The proposed models are validated on five popular
UGC VQA databases and the experimental results show that the
proposed model outperforms other state-of-the-art VQA models
by a large margin. What’s more, the proposed model trained on a
large-scale database such as LSVQ [ 44] achieves remarkable perfor-
mance when tested on the other databases without any fine-tuning,
which further demonstrates the effectiveness and generalizability
of the proposed model.
In summary, this paper makes the following contributions:
(1)We propose an effective and efficient deep learning based
model for UGC VQA, which includes the feature extraction
module, the quality regression module, and the quality pool-
ing module. The proposed model not only achieves remark-
able performance on the five popular UGC VQA databases
but also has a low computational complexity, which makes
it very suitable for practical applications.
(2)The feature extraction module extracts two kinds of quality-
aware features, the spatial features for spatial distortions and
the spatial-temporal features for motion distortions, where
the spatial features are learned from raw pixels of video
frames via an end-to-end manner and the spatial-temporal
features are extracted by a pretrained action recognition
network.
(3)We introduce a multi-scale quality fusion strategy to solve
the problem of quality assessment across different resolu-
tions, where the multi-scale weights are obtained from the
contrast sensitivity function of the human visual system by
considering the viewing environment information.
2 RELATED WORK
2.1 Handcrafted feature based NR VQA Models
A naive NR VQA method is to compute the quality of each frame
via popular NR IQA methods such as NIQE [ 24], BRISQUE [ 22],A Deep Learning based No-reference Quality Assessment Model for UGC Videos MM ’22, October 10–14, 2022, Lisboa, Portugal
Input video
Frame extraction
2D framesChunks
Global Average and STD Pooling
Spatial feature extractionStage 1Stage 2 Stage 3 Stage 4Motion feature extraction
Quality regressionQuality
ScoreMotion feature Spatial feature
Pooling3D CNN
Figure 1: The network architecture of the proposed model. The proposed model contains the feature extraction module, the
quality regression module, and the quality pooling module. The feature extraction module extracts two kinds of features, the
spatial features and the motion features.
CORNIA [ 42] etc., and then pool them into the video quality score.
A comparative study of various temporal pooling strategies on pop-
ular NR IQA methods can refer to [ 32]. The temporal information
is very important for VQA. V-BLIINDS [ 26] is a spatio-temporal
natural scene statistics (NSS) model for videos by quantifying the
NSS feature of frame-differences and motion coherency character-
istics. Mittal et al. [23] propose a training-free blind VQA model
named VIIDEO that exploits intrinsic statistics regularities of natu-
ral videos to quantify disturbances introduced due to distortions.
TLVQM [ 12] extracts abundant spatio-temporal features such as
motion, jerkiness, blurriness, noise, blockiness, color, etc. at two
levels of high and low complexity. VIDEVAL [ 33] further combines
the selected features from typical NR I/VQA methods to train a SVR
model to regress them into the video quality. Since video content
also affects its quality, especially for UGC videos, understanding the
video content is beneficial to NR VQA. Previous handcrafted feature
based methods are difficult to understand semantic information.
Hence, some studies [ 13,34] try to combine the handcrafted features
with the semantic-level features extracted by the pretrained CNN
model to improve the performance of NR VQA models. For example,
CNN-TLVQM [ 13] combines the handcrafted statistical temporal
features from TLVQM and spatial features extracted by 2D-CNN
model trained for IQA. RAPIQUE [ 34] utilizes the quality-aware
scene statistics features and semantics-aware deep CNN features
to achieve a rapid and accurate VQA model for UGC videos.
2.2 Deep learning based NR VQA Models
With the release of several large-scale VQA databases [ 8,36,44],
deep learning based NR VQA models [ 2,11,14,15,31,37,40,43,44]
attract many researchers’ attention. Liu et al. [17] propose a multi-
task BVQA model V-MEON by jointly optimizing the 3D-CNN
for quality assessment and compression distortion classification.
VSFA [ 15] first extracts the semantic features from a pre-trained
CNN model and then uses a gated recurrent unit (GRU) network
to model the temporal relationship between the semantic features
of video frames. The authors of VSFA further propose MDVSFA[16], which trains the VSFA model on the multiple VQA databases
to improve its performance and generalization. RIRNet [ 4] exploits
the effect of motion information extracted from the multi-scale
temporal frequencies for video quality assessment. Ying et al. [44]
propose a local-to-global region-based NR VQA model that com-
bines the spatial features extracted from a 2D-CNN model and the
spatial-temporal features from a 3D-CNN network. Wang et al. [37]
propose a feature-rich VQA model for UGC videos, which measures
the quality from three aspects, compression level, video content,
and distortion type and each aspect is evaluated by an individual
neural network. Xu et al. [40] first extract the spatial feature of
the video frame from a pre-trained IQA model and use the graph
convolution to extract and enhance these features, then extract
motion information from the optical flow domain, and finally inte-
grated the spatial feature and motion information via a bidirectional
long short-term memory network. Li et al. [14] also utilize the IQA
model pre-trianed on multiple databases to extract quality-aware
spatial features and the action recognition model to extract tem-
poral features, and then a GRU network is used to model spatial
and temporal features and regress them into the quality score. Wen
and Wang [ 39] propose a baseline I/VQA model for UGC videos,
which calculates the video quality by averaging the scores of each
frame and frame-level quality scores are obtained by a simple CNN
network.
3 PROPOSED MODEL
The framework of the proposed NR VQA model is shown in Fig-
ure 1, which consists of the feature extraction module, the quality
regression module, and the quality pooling module. First, we ex-
tract the quality-aware features from the spatial domain and the
spatial-temporal domain via the feature extraction module, which
are utilized to evaluate the spatial distortions and motion distor-
tions respectively. Then, the quality regression module is used to
map the quality-aware features into chunk-level quality scores. Fi-
nally, we perform the quality pooling module to obtain the video
quality score.MM ’22, October 10–14, 2022, Lisboa, Portugal Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao Zhai∗
3.1 Feature Extraction Module
In this section, we expect to extract the quality-aware features that
can represent the impact of various distortion types and content
on visual quality. The types of video distortion can be roughly
divided into two categories: the spatial distortions and the motion
distortions. The spatial distortions refer to the artifacts introduced
in the video frames, such as noise, blur, compression, low visibility,
etc. The motion distortions refer to the jitter, lagging due, etc., which
are mainly caused by unstable shooting equipment, fast-moving
objects, the low network bandwidth, etc. Therefore, we need to
extract the quality-aware features from these two aspects.
Note that the characteristics of the spatial features and motion
features are quite different. The spatial features are sensitive to
the video resolution but insensitive to the video frame rate since
the adjacent frames of the video contain lots of redundancy spatial
information and higher resolution can represent more abundant
high-frequency information, while motion features are the opposite
because the motion distortions are reflected on the temporal dimen-
sion and these features are usually consistent for local regions of
the frames.
Therefore, considering these characteristics, given a video 𝑉,
whose number of frames and frame rate are 𝑙and𝑟respectively,
we first split the video 𝑉into𝑁𝑐continuous chunks 𝑐={𝑐𝑖}𝑁𝑐
𝑖=1at
an time interval 𝜏, where𝑁𝑐=𝑙/(𝑟∗𝜏), and there are 𝑁𝑓=𝑟∗𝜏
frames in each chunk 𝑐𝑖, which is denoted as 𝑐𝑖={𝑥𝑖,𝑗}𝑁𝑓
𝑗=1. Then
we only choose a key frame 𝑥𝑖,𝑘𝑒𝑦 in each chunk to extract the
spatial features and the motion features of each chunk are extracted
using all frames in 𝑐𝑖but at a very low spatial resolution. As a
result, we can greatly reduce the computation complexity of the
VQA model with little performance degradation.
3.1.1 Spatial Feature Extraction Module. Given a frame 𝑥, we de-
note𝑓𝑤(𝑥)as the output of the CNN model 𝑓with trainable pa-
rameters𝑤={𝑤𝑘}applied on the frame 𝑥. Assume that there are
𝑁𝑠stages in the CNN model, and 𝑓𝑘𝑤(𝑥)is the output feature maps
extracted from the 𝑘-th stage, where 𝑓𝑘𝑤(𝑥)∈R𝐻𝑘×𝑊𝑘×𝐶𝑘, and𝐻𝑘,
𝑊𝑘, and𝐶𝑘are the height, width, and the number of channels of
the feature maps 𝑓𝑘𝑤(𝑥)respectively. In the following, we use the
𝑓𝑘𝑤to replace the 𝑓𝑘𝑤(𝑥)for simplicity.
It is well known that the features extracted by the deep lay-
ers of the CNN model contain rich semantic information, and are
suitable for representing content-aware features for UGC VQA.
Moreover, previous studies indicate that the features extracted by
the shallow layers of the CNN models contain low-level informa-
tion [ 28,48], which responds to low-level features such as edges,
corners, textures, etc. The low-level information is easily affected
by the distortion and is therefore distortion-aware. Hence, we ex-
tract the quality-aware features via calculating the global mean
and stand deviation of feature maps extracted from all stages of the
CNN model. Then, we apply global average and stand deviation
pooling operations on the feature maps 𝑓𝑘𝑤:
𝜇𝑓𝑘𝑤=GPavg(𝑓𝑘
𝑤),
𝜎𝑓𝑘𝑤=GPstd(𝑓𝑘
𝑤),(1)where𝜇𝑓𝑘𝑤and𝜎𝑓𝑘𝑤are the global means and stand deviation of
feature maps 𝑓𝑘𝑤respectively. Finally, we concatenate the 𝜇𝑓𝑘𝑤and
𝜎𝑓𝑘𝑤to derive the spital feature representation of our NR VQA model:
𝐹𝑘
𝑠=cat([𝜇𝑓𝑘𝑤,𝜎𝑓𝑘𝑤]),
𝐹𝑠=cat({𝐹𝑘
𝑠}𝑁𝑠
𝑘=1).(2)
3.1.2 Motion Feature Extraction Module. We extract the motion
features as the complementary quality-aware features since UGC
videos are commonly degraded by the motion distortions caused
by the unstable shooting equipment or low bit rates in the living
streaming or videoconferencing. The spatial features are difficult
to handle these distortions because they are extracted by the intra-
frames while motion distortions occur in the interframes. Therefore,
the motion features are also necessary for evaluating the quality
of UGC videos. Here, we utilize the pretrained action recognition
model as the motion feature extractor to obtain the motion features
of each video chunk. The action recognition model is designed
to detect different kinds of action classes, so the feature represen-
tation of the action recognition network can reflect the motion
information of the video to a certain extent. Therefore, given the
video chunk 𝑐and the action recognition network MOTION , we
can obtain the motion features:
𝐹𝑚=MOTION(c) (3)
where𝐹𝑚represents the motion features extract by the action
recognition network.
Therefore, given the video chunk 𝑐, we first select a key frame
in the chunk to calculate the spatial features 𝐹𝑠. Then, we calculate
the motion features 𝐹𝑚using the whole frames but at a low spatial
resolution in the video chunk. Finally, we obtain the quality-aware
features for the video chunk 𝑐by concatenating the spatial features
and motion features:
𝐹=cat([𝐹𝑠,𝐹𝑚]), (4)
3.2 Quality Regression Module
After extracting quality-aware feature representation by the feature
extraction module, we need to map these features to the quality
scores via a regression model. In this paper, we use the multi-layer
perception (MLP) as the regression model to obtain the chunk-level
quality due to its simplicity and effectiveness. The MLP consists of
two fully connected layers and there are 128 and 1 neuron in each
layer respectively. Therefore, we can obtain the chunk-level quality
score via
𝑞=𝑓𝑤FC(𝐹), (5)
where𝑓𝑤FCdenotes the function of the two FC layers and 𝑞is the
quality of the video chunk.
3.3 Quality Pooling Module
As stated in Section 3.1, we split the video 𝑉into𝑁𝑐continuous
chunks{𝑐𝑖}𝑁𝑐
𝑖=1. For the chunk 𝑐𝑖, we can obtain its chunk-level
quality score 𝑞𝑖via the feature extraction module and the quality
regression module. Then, it is necessary to pool the chunk-level
scores into the video level. Though many temporal pooling methods
have been proposed in literature [ 32][15], we find that the temporalA Deep Learning based No-reference Quality Assessment Model for UGC Videos MM ’22, October 10–14, 2022, Lisboa, Portugal
Table 1: Summary of the benchmark UGC VQA databases. Time duration: Seconds.
Database Videos Scenes Resolution Time Duration Format Distortion Type DATA Environment
KoNViD-1k [8] 1,200 1,200 540p 8 MP4 Authentic MOS + 𝜎 Crowd
YouTube-UGC [36] 1500 1500 360p-4K 20 YUV, MP4 Authentic MOS + 𝜎 Crowd
LSVQ [44] 38,811 38,811 99p-4K 5-12 MP4 Authentic MOS + 𝜎 Crowd
LBVD [3] 1,013 1,013 240p-540p 10 MP4 Authentic, Transmission MOS + 𝜎 In-lab
LIVE-YT-Gaming [45] 600 600 360p-1080p 8-9 MP4 Authentic MOS Crowd
averaging pooling achieves the best performance from Section 4.3.2.
Therefore, the video-level quality is calculated as:
𝑄=1
𝑁𝑐𝑁𝑐∑︁
𝑖=1𝑞𝑖, (6)
where𝑞𝑖is the quality of the 𝑖-th chunk and 𝑄is the video quality
evaluated by the proposed model.
3.4 Loss Function
The loss function used to optimize the proposed models consists of
two parts: the mean absolute error (MAE) loss and rank loss [ 39].
The MAE loss is used to make the evaluated quality scores close to
the ground truth, which is defined as:
𝐿𝑀𝐴𝐸=1
𝑁𝑁∑︁
𝑖=1 𝑄𝑖−ˆ𝑄𝑖 , (7)
where the ˆ𝑄𝑖is the ground truth quality score of the 𝑖-th video in a
mini-batch and 𝑁is the number of videos in the mini-batch.
The rank loss is further introduced to make the model distinguish
the relative quality of videos better, which is very useful for the
model to evaluate the videos with similar quality. Since the rank
value between two video quality is non-differentiable, we use the
following formula to approximate the rank value:
𝐿𝑖𝑗
𝑟𝑎𝑛𝑘=max(0, ˆ𝑄𝑖−ˆ𝑄𝑗 −𝑒(ˆ𝑄𝑖,ˆ𝑄𝑗)·(𝑄𝑖−𝑄𝑗)), (8)
where𝑖and𝑗are two video indexes in a mini-batch, and 𝑒(ˆ𝑄𝑖,ˆ𝑄𝑗)
is formulated as:
𝑒(ˆ𝑄𝑖,ˆ𝑄𝑗)=(
1,ˆ𝑄𝑖≥ˆ𝑄𝑗,
−1,ˆ𝑄𝑖<ˆ𝑄𝑗,(9)
Then,𝐿𝑟𝑎𝑛𝑘 is calculated via:
𝐿𝑟𝑎𝑛𝑘=1
𝑁2𝑁∑︁
𝑖=1𝑁∑︁
𝑗=1𝐿𝑖𝑗
𝑟𝑎𝑛𝑘(10)
Finally, the loss function can be obtained by:
𝐿=𝐿𝑀𝐴𝐸+𝜆·𝐿𝑟𝑎𝑛𝑘, (11)
where𝜆is a hyper-parameter to balance the MAE loss and the rank
loss.
3.5 Multi-scale Quality Fusion Strategy
Previous studies evaluate the video quality either using the original
spatial resolution or a fixed resized spatial resolution, which ig-
nore that videos are naturally multi-scale [ 53]. Some existing work
[38][25][20] shows that considering the multi-scale characteristics
can improve the performance of image quality assessment. So, wepropose a multi-scale quality fusion strategy to further improve
the evaluation accuracy of the VQA model and this strategy is
very useful to compare the quality of videos with different spatial
resolutions.
3.5.1 Multi-scale Video Quality Scores. We first resize the resolu-
tion of the video into three fixed spatial scales, which are 540p,
720p, and 1080p, respectively. We do not downscale the video from
the original scale to several lower resolution scales, which is a
more common practice in previous studies. That is because when
users watch videos in an application, the resolution of videos is
actually adapted to the resolution of the playback device, and the
modern display resolution is normally larger than 1080p. So, the
perceptual quality of the low-resolution videos is also affected by
the up-sampling artifacts, which also need to be considered by VQA
models. Therefore, given a VQA model, we can derive three qual-
ity of videos at three scales, which are denoted as 𝑄1,𝑄2, and𝑄3
respectively.
3.5.2 Adaptive Multi-scale Weights. The weight of each scale is
obtained by considering the human psychological behaviors and
the visual sensitivity characteristics. It is noted that the contrast
perception ability of the HVS depends on the spatial frequency
of the visual signal, which is modeled by the contrast sensitivity
function (CSF). Specifically, we first define a viewing resolution
factor𝜉as:
𝜉=𝜋·𝑑·𝑛
180·ℎ𝑠·2, (12)
where the unit of 𝜉is cycles per degree of visual angle (cpd), 𝑑is
the viewing distance (inch), ℎ𝑠is the height of the screen (inch),
and𝑛denotes the number of pixels in the vertical direction of the
screen. For the above three spatial scales of video, we can obtain the
corresponding 𝜉, which are denoted as 𝜉1,𝜉2, and𝜉3respectively.
We use𝜉to divide the spatial frequency range of the corresponding
scale, which covers one section of the CSF formulated by:
𝑆(𝑢)=5200𝑒(−0.0016𝑢2(1+100/𝐿)0.08)
√︃
(1+144
𝑋2
0+0.64𝑢2)(63
𝐿0.83+1
1−𝑒(−0.02𝑢2))(13)
where𝑢,𝐿, and𝑋2
0indicate spatial frequency (cpd), luminance
(cd/m2), and angular object area (squared degrees), respectively.
The weight of each scale is calculated as the area under the CSF
within the corresponding frequency covering range:
𝑤𝑖=1
𝑍∫𝜉𝑖
𝜉𝑖−1𝑆(𝑢)d𝑢,𝑖∈{1,2,3}, (14)
where𝑖from 1 to 3 corresponds the finest to coarsest scale respec-
tively, and𝜉0corresponds the viewing resolution factor of 0. 𝑍is a
normalization factor such thatÍ
𝑖𝑤𝑖=1.MM ’22, October 10–14, 2022, Lisboa, Portugal Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao Zhai∗
Table 2: Performance of the SOTA models and the proposed model on the KoNViD-1k, YouTube-UGC, LBVD, and LIVE-YT-
Gaming databases. W.A. means the weight average results. The best performing model is highlighted in each column.
TypeDatabase KoNViD-1k YouTube-UGC LBVD LIVE-YT-Gaming W.A.
Criterion SRCC PLCC SRCC PLCC SRCC PLCC SRCC PLCC SRCC PLCC
IQANIQE 0.542 0.553 0.238 0.278 0.327 0.387 0.280 0.304 0.359 0.393
BRISQUE 0.657 0.658 0.382 0.395 0.435 0.446 0.604 0.638 0.513 0.525
GM-LOG 0.658 0.664 0.368 0.392 0.314 0.304 0.312 0.317 0.433 0.440
VGG19 0.774 0.785 0.703 0.700 0.676 0.673 0.678 0.658 0.714 0.712
ResNet50 0.802 0.810 0.718 0.710 0.715 0.717 0.729 0.768 0.744 0.751
KonCept512 0.735 0.749 0.587 0.594 0.626 0.636 0.643 0.649 0.650 0.660
VQAV-BLIINDS 0.710 0.704 0.559 0.555 0.527 0.558 0.357 0.403 0.566 0.578
TLVQM 0.773 0.769 0.669 0.659 0.614 0.590 0.748 0.756 0.699 0.689
VIDEVAL 0.783 0.780 0.779 0.773 0.707 0.697 0.807 0.812 0.766 0.762
RAPIQUE 0.803 0.818 0.759 0.768 0.712 0.725 0.803 0.825 0.767 0.781
VSFA 0.773 0.775 0.724 0.743 0.622 0.642 0.776 0.801 0.721 0.736
Liel al. 0.836 0.834 0.831 0.819 - - - - - -
Pro. 0.856 0.860 0.847 0.856 0.844 0.846 0.861 0.866 0.851 0.856
Table 3: Performance of the SOTA models and the proposed
models on the LSVQ database. Pro. M.S. refers to the pro-
posed model implemented by the multi-scale quality fusion
strategy. W.A. means the weighted average results. The best
performing model is highlighted in each column.
Database Test Test-1080p W.A.
Criterion SRCC PLCC SRCC PLCC SRCC PLCC
TLVQM 0.772 0.774 0.589 0.616 0.712 0.722
VIDEVAL 0.794 0.783 0.545 0.554 0.712 0.707
VSFA 0.801 0.796 0.675 0.704 0.759 0.766
PVQ 0.827 0.828 0.711 0.739 0.789 0.799
Liel al. 0.852 0.854 0.772 0.788 0.825 0.832
Pro. 0.864 0.861 0.756 0.801 0.829 0.841
Pro. M.S. 0.867 0.861 0.764 0.803 0.833 0.842
Therefore, the multi-scale fusion quality score 𝑄𝑚is calculated
as:
𝑄𝑚=3Ö
𝑖=1𝑄𝑤𝑖
𝑖, (15)
4 EXPERIMENTAL VALIDATION
4.1 Experimental Protocol
4.1.1 Test Databases. We test the proposed model on the five UGC
VQA database: KoNViD-1k [ 8], YouTube-UGC [ 36], LSVQ [ 44],
LBVD [ 3], and LIVE-YT-Gaming [ 45]. We summarize the main infor-
mation of the databases in Table 1. The LSVQ database is the largest
UGC VQA database so far, and there are 15 video categories such
as animation, gaming, HDR, live music, sports, etc. in the YouTube-
UGC database, which is more diverse than other databases. The
LBVD database focuses on the live broadcasting videos, of which
the videos are degraded by the authentic transmission distortions.
The LIVE-YT-Gaming database consists of streamed gaming videos,
where the video content is generated by computer graphics.4.1.2 Implementation Details. We use the ResNet50 [ 7] as the back-
bone of the spatial feature extraction module and the SlowFast R50
[6] as the motion feature extraction model for the whole experi-
ments. The weights of the ResNet50 are initialized by training on
the ImageNet dataset [ 5], the weights of the SlowFast R50 are fixed
by training on the Kinetics 400 dataset [ 10], and other weights are
randomly initialized. For the spatial feature extraction module, we
resize the resolution of the minimum dimension of key frames as
520 while maintaining their aspect ratios. In the training stage, the
input frames are randomly cropped with the resolution of 448 ×448.
If we do not use the multi-scale quality fusion strategy, we crop the
center patch with the same resolutions of 448 ×448 in the testing
stage. Note that we only validate the multi-scale quality fusion
strategy on the model trained by the LSVQ database since there
are enough videos with various spatial resolutions in it. For the
motion feature extraction module, the resolution of the videos is
resized to 224×224 for both the training and testing stages. We use
PyTorch to implement the proposed models. The Adam optimizer
with the initial learning rate 0.00001 and batch size 8 are used for
training the proposed model on a server with NVIDIA V100. The
hyper-parameter 𝜆is set as 1. For simplicity, we select the first
frame in each chunk as the key frame. For the multi-scale quality
fusion strategy, there are 𝑑=35,𝑛=1080 ,ℎ=11.3,𝐿=200, and
𝑋2
0=606, and the final multi-scale weights for UGC videos are
𝑤1=0.8317,𝑤2=0.0939, and𝑤3=0.0745.
4.1.3 Comparing Algorithms. We compare the proposed method
with the following no-reference models:
•IQA models: NIQE [ 24], BRISQUE [ 22], GM-LOG [ 41], VGG19
[27], ResNet50 [7], and KonCept512 [9].
•VQA models: V-BLIINDS [ 26], TLVQM [ 12], VIDEAL [ 33],
RAPIQUE [34], VSFA [15], PVQ [44], and Li et al. [14].
Since the number of videos in the LSVQ database is relatively
large, we only compare some representative VQA models on the
LSVQ database and omit the methods which perform poorly on the
other four UGC databases.A Deep Learning based No-reference Quality Assessment Model for UGC Videos MM ’22, October 10–14, 2022, Lisboa, Portugal
4.1.4 Evaluation Criteria. We adopt two criteria to evaluate the
performance of VQA models, which are Pearson linear correlation
coefficient (PLCC) and Spearman rank-order correlation coefficient
(SRCC). PLCC reflects the prediction linearity of the VQA algorithm
and SRCC indicates the prediction monotonicity. An excellent VQA
model should obtain the value of SRCC and PLCC close to 1. Before
calculating the PLCC, we follow the same procedure in [ 1] to map
the objective score to the subject score using a four-parameter
logistic function.
For KoNViD-1k, YouTube-UGC, LBVD, and LIVE-YT-Gaming
databases, we randomly split these databases into the training set
with 80% videos and the test set with 20% videos for 10 times, and
report the median values of SRCC and PLCC. For the LSVQ database,
we follow the same training and test split suggested by [ 44] and
report the performance on the test and test-1080p subsets.
4.2 Performance Comparison with the SOTA
Models
The performance results of the VQA models on the KoNViD-1k,
YouTube-UGC, LBVD, and LIVE-YT-Gaming databases are listed in
Table 2, and on the LSVQ database are listed in Table 3. From Table
2 and Table 3, we observe that the proposed model achieves the best
performance on all five UGC VQA databases and leads by a large
margin, which demonstrates that the proposed model does have a
strong ability to measure the perceptual quality of various kinds of
UGC videos. For the test-1080p subset of the LSVQ database, the
proposed model is inferior to Li et al., which may be because the spa-
tial resolution of most videos in the test-1080p subset is larger than
1080p while the proposed model resizes the spatial resolution of test
videos into 448×448, so the proposed model has a relatively poor
ability to represent the characteristics of high-resolution videos.
Through the multi-scale quality weighting fusion strategy, the pro-
posed model can significantly improve the performance on the
test-1080p subset.
Then, most handcrafted feature based IQA models perform poorly
on these UGC VQA databases especially for the LBVD and LIVE-
YT-Gaming databases since they are designed for natural scene
images with synthetic distortions and are difficult to handle the
complex in-the-wild distortions and other video types such gaming,
videoliving, etc. It is worth noting that through fine-tuning the deep
CNN baseline i.e. ResNet50 on the VQA databases, it can achieve a
pretty good performance, which also indicates that spatial features
are very important for VQA tasks. For the NR VQA methods, the
hand-crafted feature based NR VQA methods such as TLVQM and
VIDEVAL achieve pretty well performance by incorporating the
rich spatial and temporal quality features, such as NSS features,
motion features, etc., but they are inferior to the deep learning
based NR VQA methods due to the strong feature representation
ability of CNN. VSFA extracts the spatial features from the pre-
trained image recognition model, which are not quality-aware, and
achieves relatively poor performance when compared with other
deep learning based methods. PVQ and Li et al. methods both uti-
lize the pretrained IQA model and ptretrained action recognition
model to extract spatial and motion features respectively, and they
perform better than other compared NR I/VQA methods but are
inferior to the proposed model. Through training an end-to-endTable 4: The results of ablation studies on the LSVQ data-
base. S and M means the spatial features and motion features
respectively, and S∗means that the spatial features are ex-
tracted by the pretrained image classification network.
Database Test Test-1080p
Criterion SRCC PLCC SRCC PLCC
FeatureS∗+M 0.847 0.841 0.732 0.774
S 0.827 0.829 0.702 0.757
M 0.660 0.669 0.569 0.621
RegressionGRU 0.858 0.855 0.735 0.788
Transformer 0.860 0.861 0.753 0.799
PoolingMethod in [15] 0.860 0.858 0.733 0.786
1D CNN based 0.864 0.862 0.739 0.790
Table 5: The SRCC results of cross-database evaluation. The
model is trained on the LSVQ database.
Database KoNViD-1k YouTube-UGC LBVD LIVE-YT-Gaming
Pro. 0.860 0.789 0.689 0.642
Pro. M.S. 0.859 0.822 0.711 0.683
spatial feature extractor, the proposed model can take advantage of
various video content and distortion types in the UGC databases
and learn a better quality-aware feature representation. As a result,
the proposed model achieves the best performance on all five UGC
VQA databases.
4.3 Ablation Studies
In this section, we conduct several ablation studies to investigate
the effectiveness of each module in the proposed model, including
the feature extraction module, and the quality regression module.
All the experiments are tested on the LSVQ database since it is the
largest UGC VQA model and is more representative.
4.3.1 Feature Extraction Module. The proposed model consists
of the spatial feature extractor that learns the end-to-end spatial
quality-aware features and the motion feature extractor that utilizes
a pretrained action recognition model to represent motion informa-
tion. Therefore, we first do not train the spatial feature extractor
and directly use the weights trained on the ImageNet database to
study the effect of the end-to-end training strategy for the spatial
feature extractor. Then, we only use the end-to-end trained spatial
features or the pretrained motion features to evaluate the quality of
UGC videos to investigate the effect of these two kinds of features.
The results are listed in Table 4. First, it is observed that the model
using the motion features is inferior to the model using the spatial
features and both of them are inferior to the proposed model, which
indicates that both spatial and motion features are beneficial to the
UGC VQA task and the spatial features are more important. Then,
we find that end-to-end training for the spatial feature extractor can
significantly improve the evaluation performance, which demon-
strates that end-to-end trained spatial features represent better than
that extracted by the pretrained image classification model.MM ’22, October 10–14, 2022, Lisboa, Portugal Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao Zhai∗
Table 6: Comparison of computational complexity for the six VQA models and two proposed models. Time: Second.
Methods V-BLIINDS TLVQM VIDEVAL VSFA RAPIQUE Li et al. Pro. Pro. M.S.
Time 61.982 219.992 561.408 56.424 38.126 61.971 6.929 8.448
4.3.2 Quality Regression Module. In this paper, we use the MLP
as the regression model to derive the chunk-level quality scores.
However, in previous studies, some sequential models such as GRU
[15], Transformer [ 14], etc. are also adopted to further consider the
influence of the features extracted from adjacent frames. Here, we
also adopt these methods as a comparison to investigate whether
sequential models can improve the performance of the proposed
models. Specifically, we replace the MLP module with the GRU and
Transformer and keep other experimental setups the same. The
results are listed in Table 4. We observe that models using GRU
and Transformer are both inferior to the proposed model, which
means that the MLP module is enough to regress the quality-aware
features to quality scores though it is very simple. This conclusion
is also consistent with [ 37]. The reason is that the proposed model
and the model in [ 37] calculate the chunk-level quality score and
the effect of adjacent frames are considered in the quality-aware
features (i.e. motion features), while other VQA models [ 15] [14]
calculate the frame-level quality scores, which may need to consider
the effect of adjacent frames in the quality regression module.
4.3.3 Quality Pooling Module. The proposed model uses the tem-
poral average pooling method to fuse the chunk-level quality scores
into the video level. It is noted that previous studies also propose
several temporal pooling methods for VQA. In this section, we test
two temporal pooling methods, which are the subjectively-inspired
method introduced in [ 15] and a learning based temporal pooling
method using the 1D CNN. The results are listed in Table 4. From
Table 4, we observe that the average pooling strategy achieves sim-
ilar performance to the learning based pooling method, and both of
them are superior to the subjectively-inspired methods. Since the
average pooling strategy is simpler and does not increase the extra
parameters, we use the temporal average pooling method in this
paper.
4.4 Cross-Database Evaluation
UGC videos may contain various kinds of distortions and content,
most of which may not exist in the training set. Hence, the gen-
eralization ability of the UGC VQA model is very important. In
this section, we use the cross-database evaluation to test the gener-
alization ability of the proposed model. Specifically, we train the
proposed model on the LSVQ database and test the trained model
on the other four UGC VQA databases. We list the results in Table
5. It is observed that the proposed model achieves excellent per-
formance in cross-database evaluation. The SRCC results on the
KoNViD-1k and YouTube-UGC databases both exceed 0.8, which
have surpassed most VQA models trained on the corresponding
database. We find that the multi-scale quality fusion strategy can
significantly improve the performance on the databases containing
videos with different spatial resolutions (YouTube-UGC, LBVD, and
LIVE-YT-Gaming), which further demonstrates its effectiveness.It is also observed that the performance on the LBVD and LIVE-
YT-Gaming databases is not good as the other two databases. The
reason is that the LBVD and LIVE-YT-Gaming databases contain
live broadcasting and gaming videos respectively, which may rarely
exist in the LSVQ database. Since the single database can not cover
all kinds of video types and distortions, we may further improve the
generalization ability of the proposed model via the multiple data-
base training strategy [ 30] [51] or the continual learning manner
[49] [50].
4.5 Computational Complexity
The computational complexity is a very important factor that needs
to be considered in practical applications. Hence, we test the com-
putational complexity in this section. All models are tested on a
computer with i7-6920HQ CPU, 16G RAM, and NVIDIA Quadro
P400. The deep learning based models and the handcrafted based
models are tested using the GPU and CPU respectively. We report
the running time for a video with the resolution of 1920 ×1080 and
time duration of eight seconds in Table 6. It is seen that the proposed
model has a considerably low running time compared with other
VQA models. The reason is that we use very sparse frames to calcu-
late the spatial features while other deep learning based methods
need dense frames. Moreover, we extract the motion features at a
very low resolution, which only adds little computational complex-
ity to the proposed model. The very low computational complexity
makes the proposed model suitable for practical applications.
5 CONCLUSION
In this paper, we propose an effective and efficient NR VQA model
for UGC videos. The proposed model extracts the quality-aware
features from the spatial domain and the spatial-temporal domain to
measure the spatial distortions and motion distortions respectively.
We train the spatial feature extractor in an end-to-end training
manner, so the proposed model can make full use of the various
spatial distortions and content in the current VQA database. Then,
the quality-aware features are regressed into the quality scores
by the MLP network, and the temporal average pooling is used
to obtain the video-level quality scores. We further introduce the
multi-scale quality fusion strategy to address the problem of quality
assessment across different spatial resolutions. The experimental
results show that the proposed model can effectively measure the
quality of UGC videos.
ACKNOWLEDGMENTS
This work was supported by the National Natural Science Foun-
dation of China (61831015, 61901260) and the National Key R&D
Program of China 2021YFE0206700.A Deep Learning based No-reference Quality Assessment Model for UGC Videos MM ’22, October 10–14, 2022, Lisboa, Portugal
REFERENCES
[1]Jochen Antkowiak, T Jamal Baina, France Vittorio Baroncini, Noel Chateau,
France FranceTelecom, Antonio Claudio França Pessoa, F Stephanie Colonnese,
Italy Laura Contin, Jorge Caviedes, and France Philips. 2000. Final report from
the video quality experts group on the validation of objective models of video
quality assessment march 2000. (2000).
[2]Yuqin Cao, Xiongkuo Min, Wei Sun, and Guangtao Zhai. 2021. Deep Neural Net-
works For Full-Reference And No-Reference Audio-Visual Quality Assessment.
In2021 IEEE International Conference on Image Processing (ICIP) . IEEE, 1429–1433.
[3]Pengfei Chen, Leida Li, Yipo Huang, Fengfeng Tan, and Wenjun Chen. 2019. QoE
evaluation for live broadcasting video. In 2019 IEEE International Conference on
Image Processing (ICIP) . IEEE, 454–458.
[4]Pengfei Chen, Leida Li, Lei Ma, Jinjian Wu, and Guangming Shi. 2020. RIRNet:
Recurrent-in-recurrent network for video quality assessment. In Proceedings of
the 28th ACM International Conference on Multimedia . 834–842.
[5]Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet:
A large-scale hierarchical image database. In 2009 IEEE conference on computer
vision and pattern recognition . Ieee, 248–255.
[6]Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. 2019. Slow-
fast networks for video recognition. In Proceedings of the IEEE/CVF international
conference on computer vision . 6202–6211.
[7]Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual
learning for image recognition. In Proceedings of the IEEE conference on computer
vision and pattern recognition . 770–778.
[8]Vlad Hosu, Franz Hahn, Mohsen Jenadeleh, Hanhe Lin, Hui Men, Tamás Szirányi,
Shujun Li, and Dietmar Saupe. 2017. The Konstanz natural video database
(KoNViD-1k). In 2017 Ninth international conference on quality of multimedia
experience (QoMEX) . IEEE, 1–6.
[9]Vlad Hosu, Hanhe Lin, Tamas Sziranyi, and Dietmar Saupe. 2020. KonIQ-10k: An
ecologically valid database for deep learning of blind image quality assessment.
IEEE Transactions on Image Processing 29 (2020), 4041–4056.
[10] Will Kay, Joao Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra
Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, et al .2017.
The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017).
[11] Woojae Kim, Jongyoo Kim, Sewoong Ahn, Jinwoo Kim, and Sanghoon Lee. 2018.
Deep video quality assessor: From spatio-temporal visual sensitivity to a convo-
lutional neural aggregation network. In Proceedings of the European Conference
on Computer Vision (ECCV) . 219–234.
[12] Jari Korhonen. 2019. Two-level approach for no-reference consumer video quality
assessment. IEEE Transactions on Image Processing 28, 12 (2019), 5923–5938.
[13] Jari Korhonen, Yicheng Su, and Junyong You. 2020. Blind natural video quality
prediction via statistical temporal features and deep spatial features. In Proceed-
ings of the 28th ACM International Conference on Multimedia . 3311–3319.
[14] Bowen Li, Weixia Zhang, Meng Tian, Guangtao Zhai, and Xianpei Wang. 2021.
Blindly Assess Quality of In-the-Wild Videos via Quality-aware Pre-training and
Motion Perception. arXiv preprint arXiv:2108.08505 (2021).
[15] Dingquan Li, Tingting Jiang, and Ming Jiang. 2019. Quality assessment of in-the-
wild videos. In Proceedings of the 27th ACM International Conference on Multimedia .
2351–2359.
[16] Dingquan Li, Tingting Jiang, and Ming Jiang. 2021. Unified quality assessment
of in-the-wild videos with mixed datasets training. International Journal of
Computer Vision 129, 4 (2021), 1238–1257.
[17] Wentao Liu, Zhengfang Duanmu, and Zhou Wang. 2018. End-to-End Blind
Quality Assessment of Compressed Videos Using Deep Neural Networks.. In
ACM Multimedia . 546–554.
[18] Wei Lu, Wei Sun, Xiongkuo Min, Wenhan Zhu, Quan Zhou, Jun He, Qiyuan Wang,
Zicheng Zhang, Tao Wang, and Guangtao Zhai. 2022. Deep Neural Network for
Blind Visual Quality Assessment of 4K Content. arXiv preprint arXiv:2206.04363
(2022).
[19] Pavan C Madhusudana, Xiangxu Yu, Neil Birkbeck, Yilin Wang, Balu Adsumilli,
and Alan C Bovik. 2021. Subjective and objective quality assessment of high
frame rate videos. IEEE Access 9 (2021), 108069–108082.
[20] Xiongkuo Min, Kede Ma, Ke Gu, Guangtao Zhai, Zhou Wang, and Weisi Lin.
2017. Unified blind quality assessment of compressed natural, graphic, and screen
content images. IEEE Transactions on Image Processing 26, 11 (2017), 5462–5474.
[21] Xiongkuo Min, Guangtao Zhai, Jiantao Zhou, Mylene CQ Farias, and Alan Conrad
Bovik. 2020. Study of subjective and objective quality assessment of audio-visual
signals. IEEE Transactions on Image Processing 29 (2020), 6054–6068.
[22] Anish Mittal, Anush Krishna Moorthy, and Alan Conrad Bovik. 2012. No-
reference image quality assessment in the spatial domain. IEEE Transactions on
image processing 21, 12 (2012), 4695–4708.
[23] Anish Mittal, Michele A Saad, and Alan C Bovik. 2015. A completely blind video
integrity oracle. IEEE Transactions on Image Processing 25, 1 (2015), 289–300.
[24] Anish Mittal, Rajiv Soundararajan, and Alan C Bovik. 2012. Making a “completely
blind” image quality analyzer. IEEE Signal processing letters 20, 3 (2012), 209–212.
[25] Abdul Rehman, Kai Zeng, and Zhou Wang. 2015. Display device-adapted video
quality-of-experience assessment. In Human Vision and Electronic Imaging XX ,Vol. 9394. International Society for Optics and Photonics, 939406.
[26] Michele A Saad, Alan C Bovik, and Christophe Charrier. 2014. Blind prediction
of natural video quality. IEEE Transactions on Image Processing 23, 3 (2014),
1352–1365.
[27] Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks
for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014).
[28] Wei Sun, Xiongkuo Min, Guangtao Zhai, Ke Gu, Huiyu Duan, and Siwei Ma. 2019.
MC360IQA: a multi-channel CNN for blind 360-degree image quality assessment.
IEEE Journal of Selected Topics in Signal Processing 14, 1 (2019), 64–77.
[29] Wei Sun, Xiongkuo Min, Guangtao Zhai, Ke Gu, Siwei Ma, and Xiaokang Yang.
2020. Dynamic backlight scaling considering ambient luminance for mobile
videos on lcd displays. IEEE Transactions on Mobile Computing (2020).
[30] Wei Sun, Xiongkuo Min, Guangtao Zhai, and Siwei Ma. 2021. Blind quality
assessment for in-the-wild images via hierarchical feature fusion and iterative
mixed database training. arXiv preprint arXiv:2105.14550 (2021).
[31] Wei Sun, Tao Wang, Xiongkuo Min, Fuwang Yi, and Guangtao Zhai. 2021. Deep
learning based full-reference and no-reference quality assessment models for
compressed ugc videos. In 2021 IEEE International Conference on Multimedia &
Expo Workshops (ICMEW) . IEEE, 1–6.
[32] Zhengzhong Tu, Chia-Ju Chen, Li-Heng Chen, Neil Birkbeck, Balu Adsumilli,
and Alan C Bovik. 2020. A comparative evaluation of temporal pooling methods
for blind video quality assessment. In 2020 IEEE International Conference on Image
Processing (ICIP) . IEEE, 141–145.
[33] Zhengzhong Tu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, and Alan C Bovik.
2021. UGC-VQA: Benchmarking blind video quality assessment for user generated
content. IEEE Transactions on Image Processing (2021).
[34] Zhengzhong Tu, Xiangxu Yu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, and
Alan C Bovik. 2021. Rapique: Rapid and accurate video quality prediction of user
generated content. arXiv preprint arXiv:2101.10955 (2021).
[35] Tao Wang, Zicheng Zhang, Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao
Zhai. 2022. Subjective Quality Assessment for Images Generated by Computer
Graphics. arXiv preprint arXiv:2206.05008 (2022).
[36] Yilin Wang, Sasi Inguva, and Balu Adsumilli. 2019. YouTube UGC dataset for video
compression research. In 2019 IEEE 21st International Workshop on Multimedia
Signal Processing (MMSP) . IEEE, 1–5.
[37] Yilin Wang, Junjie Ke, Hossein Talebi, Joong Gon Yim, Neil Birkbeck, Balu
Adsumilli, Peyman Milanfar, and Feng Yang. 2021. Rich features for percep-
tual quality assessment of UGC videos. In Proceedings of the IEEE/CVF Conference
on Computer Vision and Pattern Recognition . 13435–13444.
[38] Zhou Wang, Eero P Simoncelli, and Alan C Bovik. 2003. Multiscale structural sim-
ilarity for image quality assessment. In The Thrity-Seventh Asilomar Conference
on Signals, Systems & Computers, 2003 , Vol. 2. Ieee, 1398–1402.
[39] Shaoguo Wen and Junle Wang. 2021. A strong baseline for image and video
quality assessment. arXiv preprint arXiv:2111.07104 (2021).
[40] Jiahua Xu, Jing Li, Xingguang Zhou, Wei Zhou, Baichao Wang, and Zhibo Chen.
2021. Perceptual Quality Assessment of Internet Videos. In Proceedings of the
29th ACM International Conference on Multimedia . 1248–1257.
[41] Wufeng Xue, Xuanqin Mou, Lei Zhang, Alan C Bovik, and Xiangchu Feng. 2014.
Blind image quality assessment using joint statistics of gradient magnitude and
Laplacian features. IEEE Transactions on Image Processing 23, 11 (2014), 4850–
4862.
[42] Peng Ye, Jayant Kumar, Le Kang, and David Doermann. 2012. Unsupervised
feature learning framework for no-reference image quality assessment. In 2012
IEEE conference on computer vision and pattern recognition . IEEE, 1098–1105.
[43] Fuwang Yi, Mianyi Chen, Wei Sun, Xiongkuo Min, Yuan Tian, and Guangtao
Zhai. 2021. Attention Based Network For No-Reference UGC Video Quality
Assessment. In 2021 IEEE International Conference on Image Processing (ICIP) .
IEEE, 1414–1418.
[44] Zhenqiang Ying, Maniratnam Mandal, Deepti Ghadiyaram, and Alan Bovik. 2021.
Patch-VQ:’Patching Up’the Video Quality Problem. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition . 14019–14029.
[45] Xiangxu Yu, Zhenqiang Ying, Neil Birkbeck, Yilin Wang, Balu Adsumilli, and
Alan C Bovik. 2022. Subjective and Objective Analysis of Streamed Gaming
Videos. arXiv preprint arXiv:2203.12824 (2022).
[46] Saman Zadtootaghaj, Nabajeet Barman, Steven Schmidt, Maria G Martini, and
Sebastian Möller. 2018. NR-GVQM: A no reference gaming video quality metric.
In2018 IEEE International Symposium on Multimedia (ISM) . IEEE, 131–134.
[47] Saman Zadtootaghaj, Steven Schmidt, Saeed Shafiee Sabet, Sebastian Möller, and
Carsten Griwodz. 2020. Quality estimation models for gaming video streaming
services using perceptual video quality dimensions. In Proceedings of the 11th
ACM Multimedia Systems Conference . 213–224.
[48] Matthew D Zeiler and Rob Fergus. 2014. Visualizing and understanding convolu-
tional networks. In European conference on computer vision . Springer, 818–833.
[49] Weixia Zhang, Dingquan Li, Chao Ma, Guangtao Zhai, Xiaokang Yang, and Kede
Ma. 2021. Continual learning for blind image quality assessment. arXiv preprint
arXiv:2102.09717 (2021).
[50] Weixia Zhang, Kede Ma, Guangtao Zhai, and Xiaokang Yang. 2021. Task-specific
normalization for continual learning of blind image quality models. arXiv preprintMM ’22, October 10–14, 2022, Lisboa, Portugal Wei Sun, Xiongkuo Min, Wei Lu, and Guangtao Zhai∗
arXiv:2107.13429 (2021).
[51] Weixia Zhang, Kede Ma, Guangtao Zhai, and Xiaokang Yang. 2021. Uncertainty-
aware blind image quality assessment in the laboratory and wild. IEEE Transac-
tions on Image Processing 30 (2021), 3474–3486.
[52] Qi Zheng, Zhengzhong Tu, Yibo Fan, Xiaoyang Zeng, and Alan C Bovik. 2022.
No-Reference Quality Assessment of Variable Frame-Rate Videos Using Tem-
poral Bandpass Statistics. In ICASSP 2022-2022 IEEE International Conference onAcoustics, Speech and Signal Processing (ICASSP) . IEEE, 1795–1799.
[53] Qi Zheng, Zhengzhong Tu, Pavan C Madhusudana, Xiaoyang Zeng, Alan C Bovik,
and Yibo Fan. 2022. FAVER: Blind Quality Prediction of Variable Frame Rate
Videos. arXiv preprint arXiv:2201.01492 (2022).