Please use this identifier to cite or link to this item:
|Title:||Language-guided visual retrieval||Authors:||He, Su||Keywords:||Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision||Issue Date:||2021||Publisher:||Nanyang Technological University||Source:||He, S. (2021). Language-guided visual retrieval. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/151040||Abstract:||Language-guided Visual Retrieval (LGVR) is an important direction of cross-modality learning. It aims to retrieve or localize the objective message from the untrimmed visual information under the guidance of a linguistic description. In this thesis we study two popular sub-tasks of LGVR, one is Visual Grounding (VG) which aims to locate an object in the image, and the other is Natural Language Video Localization (NLVL) which aims to locate a targeted video clip from a long video span. For VG, we propose a novel modular network learning to match both the object’s symbolic feature and visual feature extracted by CNN with the linguistic information to achieve a better cross-modality alignment. Besides, a residual attention parser is raised to leverage the difficulty of understanding language expressions. For NLVL, we utilize the fine-grained semantic features of the sparse frames in the video. To organize the discrete features, we propose a network called Hybrid Graph Network to capture both the spatial and locally temporal relationships between objects in the frames and also apply semantically matching between objects and words. To model the long-span relationships between activities in the two modalities, we implement a temporal encoder based on the attentive model. Finally, we formulate the prediction as a binary classification task rather than regressing the specific boundaries. We conduct extensive experiments on popular datasets on the two tasks to validate the effectiveness of our proposed models.||URI:||https://hdl.handle.net/10356/151040||DOI:||10.32657/10356/151040||Rights:||This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0).||Fulltext Permission:||open||Fulltext Availability:||With Fulltext|
|Appears in Collections:||SCSE Theses|
Updated on May 21, 2022
Updated on May 21, 2022
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.