Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/148079
Title: | When contrastive learning meets clustering : explore inter-image contrast for image representation learning | Authors: | Li, Shenggui | Keywords: | Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence | Issue Date: | 2021 | Publisher: | Nanyang Technological University | Source: | Li, S. (2021). When contrastive learning meets clustering : explore inter-image contrast for image representation learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/148079 | Project: | SCSE20-0411 | Abstract: | Self-supervised learning has gained immense popularity in the research field of deep learning as it gets rid of the effort to label vast amounts of data. Among self-supervised learning methods, contrastive learning is a paradigm which has demonstrated high potentials in representation learning. Recent methods such as SimCLR and MoCo have delivered an impressive performance which is close to the state-of-the-art results produced by the supervised counterparts. Popular contrastive learning methods rely on instance discrimination to generate representations which are invariant after different transformations are applied. This is to explore the intra-image invariance as a single image is constrained to have similar representations when it undergoes various visual transformations and to have different representations compared to other images. However, such constraint is too strict in the sense that two different images can still look visually alike and embed similar semantics. In other words, the current methods neglect the importance of inter-image invariance as a group of similar images can also share some invariance. Thus, this project aims to explore the effect of inter-image invariance on representation learning by combining contrastive learning and clustering. Our model showed an increase in the performance in downstream tasks such as classification and outperformed the baseline models by a large margin. | URI: | https://hdl.handle.net/10356/148079 | DOI (Related Dataset): | ImageNet dataset | Schools: | School of Computer Science and Engineering | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | SCSE Student Reports (FYP/IA/PA/PI) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
FYP_Final_Report_Li_Shenggui.pdf Restricted Access | 3.14 MB | Adobe PDF | View/Open |
Page view(s)
304
Updated on Sep 15, 2024
Download(s) 50
28
Updated on Sep 15, 2024
Google ScholarTM
Check
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.