Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/144345
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Luo, Haonan | en_US |
dc.contributor.author | Lin, Guosheng | en_US |
dc.contributor.author | Liu, Zichuan | en_US |
dc.contributor.author | Liu, Fayao | en_US |
dc.contributor.author | Tang, Zhenmin | en_US |
dc.contributor.author | Yao, Yazhou | en_US |
dc.date.accessioned | 2020-10-29T06:02:35Z | - |
dc.date.available | 2020-10-29T06:02:35Z | - |
dc.date.issued | 2019 | - |
dc.identifier.citation | Luo, H., Lin, G., Liu, Z., Liu, F., Tang, Z., & Yao, Y. (2019). SegEQA : video segmentation based visual attention for embodied question answering. Proceedings of the International Conference on Computer Vision (ICCV) 2019. doi:10.1109/ICCV.2019.00976 | en_US |
dc.identifier.uri | https://hdl.handle.net/10356/144345 | - |
dc.description.abstract | Embodied Question Answering (EQA) is a newly defined research area where an agent is required to answer the user's questions by exploring the real world environment. It has attracted increasing research interests due to its broad applications in automatic driving system, in-home robots, and personal assistants. Most of the existing methods perform poorly in terms of answering and navigation accuracy due to the absence of local details and vulnerability to the ambiguity caused by complicated vision conditions. To tackle these problems, we propose a segmentation based visual attention mechanism for Embodied Question Answering. Firstly, We extract the local semantic features by introducing a novel high-speed video segmentation framework. Then by the guide of extracted semantic features, a bottom-up visual attention mechanism is proposed for the Visual Question Answering (VQA) sub-task. Further, a feature fusion strategy is proposed to guide the training of the navigator without much additional computational cost. The ablation experiments show that our method boosts the performance of VQA module by 4.2% (68.99% vs 64.73%) and leads to 3.6% (48.59% vs 44.98%) overall improvement in EQA accuracy. | en_US |
dc.description.sponsorship | AI Singapore | en_US |
dc.description.sponsorship | Ministry of Education (MOE) | en_US |
dc.description.sponsorship | National Research Foundation (NRF) | en_US |
dc.language.iso | en | en_US |
dc.relation | AISG-RP-2018-003 | en_US |
dc.relation | RG126/17 (S) | en_US |
dc.rights | © 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/ICCV.2019.00976 | en_US |
dc.subject | Engineering::Computer science and engineering | en_US |
dc.title | SegEQA : video segmentation based visual attention for embodied question answering | en_US |
dc.type | Conference Paper | en |
dc.contributor.school | School of Computer Science and Engineering | en_US |
dc.contributor.conference | International Conference on Computer Vision (ICCV) 2019 | en_US |
dc.identifier.doi | 10.1109/ICCV.2019.00976 | - |
dc.description.version | Accepted version | en_US |
dc.subject.keywords | Computer Vision | en_US |
dc.subject.keywords | Image Fusion | en_US |
dc.citation.conferencelocation | Seoul, Korea (South) | en_US |
dc.description.acknowledgement | The authors would like to thank the financial support from the program of China Scholarships Council (No.201806840059). This work is partly supported by the National Research Foundation Singapore under its AI Singapore Programme [AISG-RP-2018-003] and the MOE Tier-I research grant [RG126/17 (S)]. We would like to thank NVIDIA for GPU donation. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not reflect the views of National Research Foundation, Singapore. | en_US |
item.grantfulltext | open | - |
item.fulltext | With Fulltext | - |
Appears in Collections: | SCSE Conference Papers |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
gusoheng paper4 iccv 2019.pdf | 3.67 MB | Adobe PDF | ![]() View/Open |
SCOPUSTM
Citations
20
18
Updated on May 25, 2023
Web of ScienceTM
Citations
20
14
Updated on May 28, 2023
Page view(s)
259
Updated on Jun 2, 2023
Download(s) 20
194
Updated on Jun 2, 2023
Google ScholarTM
Check
Altmetric
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.