Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/138421
Title: Vision-based navigation of an unmanned surface vehicle with object detection and tracking abilities
Authors: Shin, Bok-Suk
Mou, Xiaozheng
Mou, Wei
Wang, Han
Keywords: Engineering::Electrical and electronic engineering
Issue Date: 2017
Source: Shin, B.-S., Mou, X., Mou, W., & Wang, H. (2018). Vision-based navigation of an unmanned surface vehicle with object detection and tracking abilities. Machine Vision and Applications, 29, 95-112. doi:10.1007/s00138-017-0878-7
Journal: Machine Vision and Applications
Abstract: The paper discusses autocalibration, object detection, and object tracking for unmanned surface vehicles. Input data are recorded with a wide-baseline stereo vision system providing accuracy for distance estimations. The paper reports about followed ways and novel contributions for ensuring a working system solution. Automatic self-calibration is used for the wide-baseline stereo vision system. Robust sea surface estimation and the detection of the horizon support the understanding of the given scene environment. Long-range (i.e. up to 500 m) object detection and tracking are supported by the used wide-baseline stereo system. The paper informs about the complete system design, informs about applied or designed methods, and also about experiments which verify that the system achieved an operational state.
URI: https://hdl.handle.net/10356/138421
ISSN: 0932-8092
DOI: 10.1007/s00138-017-0878-7
Rights: © 2017 Springer-Verlag GmbH Germany. All rights reserved. This paper was published in Machine Vision and Applications and is made available with permission of Springer-Verlag GmbH Germany.
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:EEE Journal Articles

SCOPUSTM   
Citations 10

17
Updated on Mar 10, 2021

PublonsTM
Citations 20

14
Updated on Mar 9, 2021

Page view(s)

30
Updated on May 7, 2021

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.