Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/181046
Title: Blind video quality prediction by uncovering human video perceptual representation
Authors: Liao, Liang
Xu, Kangmin
Wu, Haoning
Chen, Chaofeng
Sun, Wenxiu
Yan, Qiong
Kuo, Jay C.-C.
Lin, Weisi
Keywords: Computer and Information Science
Issue Date: 2024
Source: Liao, L., Xu, K., Wu, H., Chen, C., Sun, W., Yan, Q., Kuo, J. C. & Lin, W. (2024). Blind video quality prediction by uncovering human video perceptual representation. IEEE Transactions On Image Processing, 33, 4998-5013. https://dx.doi.org/10.1109/TIP.2024.3445738
Project: IAF-ICP 
Journal: IEEE Transactions on Image Processing 
Abstract: Blind video quality assessment (VQA) has become an increasingly demanding problem in automatically assessing the quality of ever-growing in-the-wild videos. Although efforts have been made to measure temporal distortions, the core to distinguish between VQA and image quality assessment (IQA), the lack of modeling of how the human visual system (HVS) relates to the temporal quality of videos hinders the precise mapping of predicted temporal scores to the human perception. Inspired by the recent discovery of the temporal straightness law of natural videos in the HVS, this paper intends to model the complex temporal distortions of in-the-wild videos in a simple and uniform representation by describing the geometric properties of videos in the visual perceptual domain. A novel videolet, with perceptual representation embedding of a few consecutive frames, is designed as the basic quality measurement unit to quantify temporal distortions by measuring the angular and linear displacements from the straightness law. By combining the predicted score on each videolet, a perceptually temporal quality evaluator (PTQE) is formed to measure the temporal quality of the entire video. Experimental results demonstrate that the perceptual representation in the HVS is an efficient way of predicting subjective temporal quality. Moreover, when combined with spatial quality metrics, PTQE achieves top performance over popular in-the-wild video datasets. More importantly, PTQE requires no additional information beyond the video being assessed, making it applicable to any dataset without parameter tuning. Additionally, the generalizability of PTQE is evaluated on video frame interpolation tasks, demonstrating its potential to benefit temporal-related enhancement tasks.
URI: https://hdl.handle.net/10356/181046
ISSN: 1941-0042
DOI: 10.1109/TIP.2024.3445738
Schools: School of Computer Science and Engineering 
Rights: © 2024 IEEE. All rights reserved.
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:SCSE Journal Articles

Page view(s)

45
Updated on Jan 16, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.