dc.contributor.authorHu, Yiqun
dc.date.accessioned2009-07-20T07:37:46Z
dc.date.accessioned2017-07-23T08:28:47Z
dc.date.available2009-07-20T07:37:46Z
dc.date.available2017-07-23T08:28:47Z
dc.date.copyright2008en_US
dc.date.issued2008
dc.identifier.citationHu, Y. Q. (2008). Detection of visual attention regions in images and videos. Doctoral thesis, Nanyang Technological University, Singapore.
dc.identifier.urihttp://hdl.handle.net/10356/18864
dc.description.abstractThe explosive growth of multimedia content and advances in the development of hardware with multimedia functionalities call for techniques to enable users to access such content anywhere and anytime and with similarly pleasing experience each time. This requires intelligent search, transmission, analysis and display of multimedia data. However, in addition to the data being very large in size, it is inherently complex due to the variety of features (color, texture, shapes, motion, etc.) that it contains. The challenge then is to detect information front the clutter for further processing. The relevant information is the visual attention region (VAR) whose detection in images and videos is the topic of this dissertation. The bottom-up model for detecting VAR in an image involves generation of a saliency map that highlights contrasts in features like color, intensity and orientation. The saliency map itself is obtained through a combination of each feature map that highlights the contrast for that particular feature. We investigate the process of good selection and proper combination strategies for the features. We propose a novel Composite Saliency Indicator (CSI) to determine the contribution of each feature map to the salient region. CS1 is designed to capture the spatial compactness as well as the density of candidate regions in the feature maps. We also propose a Context Suppression Model that provides a measure to determine similarity among candidate attention regions in a feature map. This measure is used to find a suppression factor for a particular patch in the scene, which is then used to highlight actual attention regions. We also demonstrate an application in multimedia adaptation that benefits from the improved VAR detection.en_US
dc.format.extent163 p.en_US
dc.language.isoenen_US
dc.subjectDRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer visionen_US
dc.titleDetection of visual attention regions in images and videosen_US
dc.typeThesis
dc.contributor.schoolSchool of Computer Engineeringen_US
dc.contributor.supervisorChia Liang Tienen_US
dc.contributor.supervisorDeepu Rajanen_US
dc.description.degreeDOCTOR OF PHILOSOPHY (SCE)en_US


Files in this item

FilesSizeFormatView
HuYiQun08.pdf25.16Mbapplication/pdfView/Open

This item appears in the following Collection(s)

Show simple item record