Please use this identifier to cite or link to this item:
|Title:||Sparse visual signal representations and selected applications||Authors:||Hung, Tzu-Yi||Keywords:||DRNTU::Engineering::Electrical and electronic engineering::Computer hardware, software and systems||Issue Date:||2015||Source:||Hung, T.-Y. (2015). Sparse visual signal representations and selected applications. Doctoral thesis, Nanyang Technological University, Singapore.||Abstract:||Sparse representation has been well investigated and discussed over the past decade due to its ability in visual signal discrimination for various applications such as face recognition, image classification and video clustering. It has attracted more and more interest in the recent years because of the increasing demands for developing real world systems with large-scale image and video collections. While a large number of sparse representation algorithms have been proposed in the literature and some encouraging results have been obtained, there is still a need for further improvement. This thesis aims to address various issues of sparse representation, including feature quantization models, sparsity estimation methods and dictionary learning techniques for sparse visual signal representation over different computer vision and pattern recognition tasks such as image classification, action recognition and activity-based human identification to demonstrate their efficacy and superiority over state-of-the-art methods. More specifically, we focus our work on two directions: 1) An application-oriented problem: we investigate the problem of activity-based person identification which will be elaborated in the thesis; and 2) A model-oriented problem: we improve the existing sparse coding approaches in a more efficient and effective way and evaluate the performance of the proposed method on several visual tasks.||URI:||http://hdl.handle.net/10356/65048||Fulltext Permission:||open||Fulltext Availability:||With Fulltext|
|Appears in Collections:||EEE Theses|
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.