Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/140433
Title: Deep learning for object detection and image segmentation
Authors: Tan, Yan Hwa
Keywords: Engineering::Mechanical engineering
Issue Date: 2020
Publisher: Nanyang Technological University
Project: C086
Abstract: In recent years, the fast-moving consumer goods (FMCG) industry has shown significant interest in robot warehouse automation technology due to the increasing demand of e-commerce, fast and reliable delivery. However, it is not a simple task to pack a large variety of products according to mass customized orders. Therefore, a fully autonomous warehouse pick-and-place system is able to complete the job with ease by employing a robust vision system that reliably locates and recognizes objects from cluttered environment, different objects and self-occlusions. The aim of this project is to develop an automated solution to allow the robot to pick up the indicated object accurately from a clustered bin in bin-picking. The robot system setup consists of a UR5 robotic arm attached with a gripper and a vision camera. In the proposed approach, we segmented and labelled multiple perspective of a view using a convolutional neural network. A large amount of training data is required to train a deep neural network for segmentation. Therefore, the proposed solution used a self-supervised method to train a large dataset and at a faster speed. The Mask-R-CNN approach was also implemented to identify each item and their individual masks to achieve a higher accuracy for object detection and image segmentation.
URI: https://hdl.handle.net/10356/140433
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:MAE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FINAL REPORT_Tan Yan Hwa.pdf
  Restricted Access
5.9 MBAdobe PDFView/Open

Page view(s)

183
Updated on Dec 3, 2022

Download(s) 50

30
Updated on Dec 3, 2022

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.