Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/170015
Title: | Gesture recognition based on deep learning | Authors: | Yang, Chaoran | Keywords: | Engineering::Computer science and engineering | Issue Date: | 2023 | Publisher: | Nanyang Technological University | Source: | Yang, C. (2023). Gesture recognition based on deep learning. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/170015 | Abstract: | Gesture recognition based on deep learning is a rapidly growing field of research and development that has the potential to revolutionize the way humans interact with computers and machines. Gesture recognition involves using algorithms and techniques to interpret human gestures, such as hand and body movements, facial expressions, and vocal intonations, to understand their meaning and intent. Gesture recognition based on deep learning has a wide range of potential applications in fields such as robotics, human-computer interaction, and healthcare. The aim of this dissertation is to design gesture recognition algorithm based on deep learning. Object detection based on convolutional neural network can be divided into one-stage object detection and two-stage object detection. Firstly, this dissertation investigates and introduces literature review, and then based on YOLOV3 and RESNET-50, this dissertation gives the models of two detection methods and the training and testing results of these two models on the dataset. | URI: | https://hdl.handle.net/10356/170015 | Schools: | School of Electrical and Electronic Engineering | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | EEE Theses |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Gesture Recognition based on Deep Learning.pdf Restricted Access | 16.4 MB | Adobe PDF | View/Open |
Page view(s)
232
Updated on May 7, 2025
Download(s)
15
Updated on May 7, 2025
Google ScholarTM
Check
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.