Please use this identifier to cite or link to this item:
Title: A supervised two-channel learning method for hidden Markov model and application on lip reading
Authors: Foo, Say Wei
Dong, Liang
Issue Date: 2002
Source: Foo, S. W., & Dong, L. (2002). A supervised two-channel learning method for hidden Markov model and application on lip reading. IEEE International Conference on Advanced Learning Technologies.
Abstract: In this paper, a novel two-channel learning method for hidden Markov model (HMM) is proposed. This method is specially designed to train HMMs for fine recognition from similar observations. The prominent features of this method are 1.) the criterion function is based on the difference between training sequences, and 2.) a twochannel structure is adopted to maintain the validity of the HMM. This learning method has been applied on a viseme-level lip reading system. The result shows that the performance of the two channel approach is better than that of the maximum likelihood (ML) estimation.
Rights: © IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.
Fulltext Permission: open
Fulltext Availability: With Fulltext
Appears in Collections:EEE Conference Papers

Files in This Item:
File Description SizeFormat 
C60S-ICALT2002DongL.pdfAccepted version181.37 kBAdobe PDFThumbnail

Google ScholarTM


Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.