Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/61627
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSelvaraj, Prabu
dc.date.accessioned2014-06-30T04:29:35Z
dc.date.available2014-06-30T04:29:35Z
dc.date.copyright2013en_US
dc.date.issued2013
dc.identifier.urihttp://hdl.handle.net/10356/61627
dc.description.abstractLocalization is the process of making a target or robot realize its current location. It is one of the most important task in autonomous navigation of either UGV (Unmanned Ground Vehicle) or UAV (Unmanned Aerial Vehicle). Localization is the basic information for any robot navigation tasks. Without localization information, robot's path cannot be tracked. Many robotic tasks such as obstacle avoidance, path planning, indoor and outdoor exploration etc ..., relies on position information of the robot. Localization is performed by tracking robot's position throughout the robot's movement using sensors. Most researchers use Global Positioning System (GPS) as the sensor to directly obtain the robot's location. But GPS signal can be interrupted by tall buildings, dense trees etc…In those cases, we need to rely on other sensor such as Laser, IMU (Inertial Measurement Units), Camera etc…for localization. In this research work, a vision based localization technique is proposed. The significance of this vision based localization method is that it can guide the robot more reliably even in the absence of GPS signal. A Bumblebee 2 stereo camera is used in this research to localize a robot with suitable computer vision techniques. The novelty of the algorithm is that two algorithms are developed in which one is used for robust pose estimate and other is used for global localization. In the entire experiments, we use stereo camera as the only sensor. Experiments were performed using commercially available Pioneer robot as well as using few custom made robots in the lab. Suitable feature extraction and feature tracking techniques were used for obtaining the robot's current pose. Scale Invariant Feature Transform (SIFT) is used as the feature detector. These features are tracked in subsequent frames. The initial and final position of the features are calcu-lated from which rotation and translation are estimated. From the estimated rotation and translation, localization is achieved. Finally few techniques to achieve localization using single camera are discussed and methods for improving existing localization techniques are also suggested.en_US
dc.format.extent94 p.en_US
dc.language.isoenen_US
dc.subjectDRNTU::Engineering::Electrical and electronic engineeringen_US
dc.titleVision based localization of a moving targeten_US
dc.typeThesis
dc.contributor.schoolSchool of Electrical and Electronic Engineeringen_US
dc.description.degreeMaster of Science (Computer Control and Automation)en_US
dc.contributor.supervisor2Hu Quoqiangen_US
item.grantfulltextrestricted-
item.fulltextWith Fulltext-
Appears in Collections:EEE Theses
Files in This Item:
File Description SizeFormat 
G1201862A.pdf
  Restricted Access
9.45 MBAdobe PDFView/Open

Page view(s)

323
Updated on Jul 22, 2024

Download(s)

16
Updated on Jul 22, 2024

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.