Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/148079
Title: When contrastive learning meets clustering : explore inter-image contrast for image representation learning
Authors: Li, Shenggui
Keywords: Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Issue Date: 2021
Publisher: Nanyang Technological University
Source: Li, S. (2021). When contrastive learning meets clustering : explore inter-image contrast for image representation learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/148079
Project: SCSE20-0411
Abstract: Self-supervised learning has gained immense popularity in the research field of deep learning as it gets rid of the effort to label vast amounts of data. Among self-supervised learning methods, contrastive learning is a paradigm which has demonstrated high potentials in representation learning. Recent methods such as SimCLR and MoCo have delivered an impressive performance which is close to the state-of-the-art results produced by the supervised counterparts. Popular contrastive learning methods rely on instance discrimination to generate representations which are invariant after different transformations are applied. This is to explore the intra-image invariance as a single image is constrained to have similar representations when it undergoes various visual transformations and to have different representations compared to other images. However, such constraint is too strict in the sense that two different images can still look visually alike and embed similar semantics. In other words, the current methods neglect the importance of inter-image invariance as a group of similar images can also share some invariance. Thus, this project aims to explore the effect of inter-image invariance on representation learning by combining contrastive learning and clustering. Our model showed an increase in the performance in downstream tasks such as classification and outperformed the baseline models by a large margin.
URI: https://hdl.handle.net/10356/148079
DOI (Related Dataset): ImageNet dataset
Schools: School of Computer Science and Engineering 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FYP_Final_Report_Li_Shenggui.pdf
  Restricted Access
3.14 MBAdobe PDFView/Open

Page view(s)

304
Updated on Sep 15, 2024

Download(s) 50

28
Updated on Sep 15, 2024

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.