Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/173445
Title: Neighbourhood representative sampling for efficient end-to-end video quality assessment
Authors: Wu, Haoning
Chen, Chaofeng
Liao, Liang
Hou, Jingwen
Sun, Wenxiu
Yan, Qiong
Gu, Jinwei
Lin, Weisi
Keywords: Computer and Information Science
Issue Date: 2023
Source: Wu, H., Chen, C., Liao, L., Hou, J., Sun, W., Yan, Q., Gu, J. & Lin, W. (2023). Neighbourhood representative sampling for efficient end-to-end video quality assessment. IEEE Transactions On Pattern Analysis and Machine Intelligence, 45(12), 15185-15202. https://dx.doi.org/10.1109/TPAMI.2023.3319332
Journal: IEEE Transactions on Pattern Analysis and Machine Intelligence
Abstract: The increased resolution of real-world videos presents a dilemma between efficiency and accuracy for deep Video Quality Assessment (VQA). On the one hand, keeping the original resolution will lead to unacceptable computational costs. On the other hand, existing practices, such as resizing or cropping, will change the quality of original videos due to difference in details or loss of contents, and are henceforth harmful to quality assessment. With obtained insight from the studies of spatial-temporal redundancy in the human visual system, visual quality around a neighbourhood has high probability to be similar, and this motivates us to investigate an effective quality-sensitive neighbourhood representative sampling scheme for VQA. In this work, we propose a unified scheme, spatial-temporal grid mini-cube sampling (St-GMS), and the resultant samples are named fragments. In St-GMS, full-resolution videos are first divided into mini-cubes with predefined spatial-temporal grids, then the temporal-aligned quality representatives are sampled to compose the fragments that serve as inputs for VQA. In addition, we design the Fragment Attention Network (FANet), a network architecture tailored specifically for fragments. With fragments and FANet, the proposed FAST-VQA and FasterVQA (with an improved sampling scheme) achieves up to 1612× efficiency than the existing state-of-the-art, meanwhile achieving significantly better performance on all relevant VQA benchmarks.
URI: https://hdl.handle.net/10356/173445
ISSN: 0162-8828
DOI: 10.1109/TPAMI.2023.3319332
Schools: School of Computer Science and Engineering 
Research Centres: S-Lab
Rights: © 2023 IEEE. All rights reserved.
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:SCSE Journal Articles

SCOPUSTM   
Citations 20

28
Updated on Mar 13, 2025

Page view(s)

92
Updated on Mar 15, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.