Please use this identifier to cite or link to this item:
Title: Deep-learning-based 3D driver pose estimation for autonomous driving
Authors: Cao, Xiao
Keywords: Engineering::Computer science and engineering
Engineering::Mechanical engineering::Mechatronics
Issue Date: 2021
Publisher: Nanyang Technological University
Source: Cao, X. (2021). Deep-learning-based 3D driver pose estimation for autonomous driving. Master's thesis, Nanyang Technological University, Singapore.
Abstract: Human-machine interaction is a key for the future development of virtual reality, augmented reality, artificial intelligence and smart device. The application of human-machine interaction technology, especially human body estimation, in autonomous driving is important to facilitate drivers to drive safely and smoothly. Human estimation can detect driver fatigue. It can also help ergonomics research and then improve human-machine interface design in automated vehicles. Researchers have got great achievements in human state estimation, including body estimation, hand estimation and face estimation. In the past, human estimation technology is dependent on hardware devices while estimation methods based on machine learning and deep learning become increasingly more popular and show excellent performance compared with traditional ways in terms of cost and efficiency. However, most estimation models are developed separately, which means that the existing models can only process body estimation or hand estimation separately instead simultaneously, while the model that can identify different parts of human at the same time is more expected in the research and application. In this dissertation, five deep learning models, including Simple Faster R-CNN, RootNet, PoseNet, YOLOv3 and a hand estimation model, are selected and then combined through a cascade method to develop an integrated model which can estimate the human body and human hand simultaneously. The outputs of each model are saved in different coordinate systems, so they cannot be fed into the subsequent neural network directly. Hence, in this project, they are transformed into the same coordinate system by a rotation transformation matrix and that enables five models to be connected in series. Through the experiment designed specifically, the integrated model is proven to be able to produce 2D pose and 3D pose of the human body and human hands at the same time. In this project, many problems still exist. These problems will be solved and other functions, such as the face estimation models will be added in the future.
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:MAE Theses

Files in This Item:
File Description SizeFormat 
  Restricted Access
3.14 MBAdobe PDFView/Open

Page view(s)

Updated on Jan 20, 2022


Updated on Jan 20, 2022

Google ScholarTM


Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.