Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/74095
Title: Graph convolutional neural networks for text categorization
Authors: Lakhotia, Suyash
Keywords: DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Issue Date: 2018
Abstract: Text categorization is the task of labelling text data from a predetermined set of thematic labels. In recent years, it has become of increasing importance as we generate large volumes of data and require the ability to search through these vast datasets with flexible queries. However, manually labelling text data is an extremely tedious task that is prone to human error. Thus, text classification has become a key focus of machine learning research, with the goal of producing models that are more efficient and accurate than traditional methods. This project explores the recently enhanced deep learning techniques of convolutional neural networks and their fusion with graph analysis (i.e. graph convolutional neural networks) in the field of text categorization and compares their performance to established baseline models and simpler multilayer perceptrons. We show through experiments on three major text classification datasets (Rotten Tomatoes Sentence Polarity, 20 Newsgroups and Reuters Corpus Volume 1) that graph convolutional neural networks can naturally work in the space of words represented as a graph and perform with greater or similar test accuracy when compared to standard convolutional neural networks and simpler baseline models.
URI: http://hdl.handle.net/10356/74095
Rights: Nanyang Technological University
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
Final Year Project Report.pdf
  Restricted Access
2.53 MBAdobe PDFView/Open

Page view(s) 50

265
Updated on Nov 29, 2020

Download(s)

73
Updated on Nov 29, 2020

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.