Please use this identifier to cite or link to this item:
Title: Implementation of fast SLAM on a UAV
Authors: Muhammad Lutfan Mikail Yang Razali
Keywords: DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
Issue Date: 2018
Abstract: Previous Simultaneous Localization and Mapping (SLAM) methods are time-consuming iterative algorithms. A PhD student whom I am working with for this project has developed a non-iterative algorithm that produces a closed-form solution to this SLAM problem. This algorithm works with a O(n lg n) time complexity and is faster than traditional SLAM algorithms. Despite that, the non-iterative SLAM algorithm developed by the PhD student is still not efficient enough to be used optimally on Unmanned Aerial Vehicles (UAVs). This is due to the fact that UAVs require a light body for stable flight. The light weight can only be achieved by using lighter hardware. However, lighter hardware often means that they are less computationally powerful. As a result, the aim of this study was to optimize the existing non-iterative SLAM algorithm via parallel programming such that it can run optimally on incredibly low-power CPUs. This study has managed to design and restructure a non-iterative SLAM algorithm which is 34% faster than the original --- allowing the algorithm to work as efficiently on low-power CPUs as it does on computationally powerful CPUs. This allows UAVs to run the SLAM algorithm on lightweight, low-power CPUs and hence, achieve more stable flight due to the lack of heavy hardware.
Schools: School of Electrical and Electronic Engineering 
Rights: Nanyang Technological University
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:EEE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
  Restricted Access
Report for FYP2.38 MBAdobe PDFView/Open

Page view(s)

Updated on Jul 13, 2024

Download(s) 50

Updated on Jul 13, 2024

Google ScholarTM


Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.