Please use this identifier to cite or link to this item:
|Title:||SVD-based quality metric for image and video using machine learning||Authors:||Narwaria, Manish
|Keywords:||DRNTU::Engineering::Computer science and engineering||Issue Date:||2011||Source:||Narwaria, M., & Lin, W. (2012). SVD-Based Quality Metric for Image and Video Using Machine Learning. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 42(2), 347-364.||Series/Report no.:||IEEE transactions on systems, man, and cybernetics, part b (cybernetics)||Abstract:||We study the use of machine learning for visual quality evaluation with comprehensive singular value decomposition (SVD)-based visual features. In this paper, the two-stage process and the relevant work in the existing visual quality metrics are first introduced followed by an in-depth analysis of SVD for visual quality assessment. Singular values and vectors form the selected features for visual quality assessment. Machine learning is then used for the feature pooling process and demonstrated to be effective. This is to address the limitations of the existing pooling techniques, like simple summation, averaging, Minkowski summation, etc., which tend to be ad hoc. We advocate machine learning for feature pooling because it is more systematic and data driven. The experiments show that the proposed method outperforms the eight existing relevant schemes. Extensive analysis and cross validation are performed with ten publicly available databases (eight for images with a total of 4042 test images and two for video with a total of 228 videos). We use all publicly accessible software and databases in this study, as well as making our own software public, to facilitate comparison in future research.||URI:||https://hdl.handle.net/10356/96456
|ISSN:||1083-4419||DOI:||http://dx.doi.org/10.1109/TSMCB.2011.2163391||Rights:||© 2011 IEEE.||Fulltext Permission:||none||Fulltext Availability:||No Fulltext|
|Appears in Collections:||SCSE Journal Articles|
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.