Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/143110
Title: | ConvNet-based visual place recognition under appearance changes for unmanned vehicles | Authors: | Li, Heshan | Keywords: | Engineering::Electrical and electronic engineering | Issue Date: | 2020 | Publisher: | Nanyang Technological University | Abstract: | Identifying the place unmanned vehicles have visited is crucial for their re-localization to eliminate accumulating drifts. As a VPR(visual place recognition problem), the goal is to retrieve the correct reference frames in database which depict the same place as given query image.For retrieval-basedVPR, methods can be classified as VLAD-based and sum-based. In this dissertation, I present the following contributions. FirstlyI reproduced pipeline of the algorithm, and then trained the modelswhose backbone are alexnet or VGG16 and head architecture are max, avg or VLAD-based pooling layer NetVLAD respectively on the Pittsburgh 30k training setand test them on Pittsburgh 120k test and Pittsburgh 30k val. Then, I evaluatedtheirabilitiesof generalizationby applying them on the revisited image retrieval testing datasets roxford5k and rparis6k. And byintroducing indicator mAP and mP@, their overall performancesarebetter evaluated and compared. What’s more, I reproduced another sum-basedpooling layer APANet, then trained and evaluatedits performance. Finally I showedthat NetVLAD possesses the overall best performance, APANet enjoys greater improvement compared to the sumpooling. | URI: | https://hdl.handle.net/10356/143110 | Schools: | School of Electrical and Electronic Engineering | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | EEE Theses |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Final Dissertation_Li_Heshan.pdf Restricted Access | 4.25 MB | Adobe PDF | View/Open |
Page view(s)
394
Updated on May 7, 2025
Download(s)
21
Updated on May 7, 2025
Google ScholarTM
Check
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.