Please use this identifier to cite or link to this item:
Title: Scalable image quality assessment with 2D mel-cepstrum and machine learning approach
Authors: Narwaria, Manish
Lin, Weisi
Cetin, A. Enis
Keywords: DRNTU::Engineering::Computer science and engineering
Issue Date: 2012
Source: Narwaria, M., Lin, W., & Cetin, A. E. (2012). Scalable image quality assessment with 2D mel-cepstrum and machine learning approach. Pattern recognition, 45(1), 299-313.
Series/Report no.: Pattern recognition
Abstract: Measurement of image quality is of fundamental importance to numerous image and video processing applications. Objective image quality assessment (IQA) is a two-stage process comprising of the following: (a) extraction of important information and discarding the redundant one, (b) pooling the detected features using appropriate weights. These two stages are not easy to tackle due to the complex nature of the human visual system (HVS). In this paper, we first investigate image features based on two-dimensional (2D) mel-cepstrum for the purpose of IQA. It is shown that these features are effective since they can represent the structural information, which is crucial for IQA. Moreover, they are also beneficial in a reduced-reference scenario where only partial reference image information is used for quality assessment. We address the second issue by exploiting machine learning. In our opinion, the well established methodology of machine learning/pattern recognition has not been adequately used for IQA so far; we believe that it will be an effective tool for feature pooling since the required weights/parameters can be determined in a more convincing way via training with the ground truth obtained according to subjective scores. This helps to overcome the limitations of the existing pooling methods, which tend to be over simplistic and lack theoretical justification. Therefore, we propose a new metric by formulating IQA as a pattern recognition problem. Extensive experiments conducted using six publicly available image databases (totally 3211 images with diverse distortions) and one video database (with 78 video sequences) demonstrate the effectiveness and efficiency of the proposed metric, in comparison with seven relevant existing metrics.
ISSN: 0031-3203
DOI: 10.1016/j.patcog.2011.06.023
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:SCSE Journal Articles

Citations 10

Updated on Jul 16, 2020

Citations 10

Updated on Mar 9, 2021

Page view(s) 20

Updated on Apr 13, 2021

Google ScholarTM




Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.