Please use this identifier to cite or link to this item:
Title: Semi-supervised generative open information extraction
Authors: Zhou, Shaowen
Keywords: Engineering::Computer science and engineering
Issue Date: 2023
Publisher: Nanyang Technological University
Source: Zhou, S. (2023). Semi-supervised generative open information extraction. Master's thesis, Nanyang Technological University, Singapore.
Abstract: Open Information Extraction (OpenIE) extracts facts in the form of n-ary relation tuples, i.e., (arg1, predicate, arg2, …, argn), from unstructured text without relying on predefined ontology schema. It has the potential to handle heterogeneous corpora with minimal human intervention. With OpenIE, Web-scale unconstrained IE systems can be developed to acquire large quantities of knowledge. The knowledge gathered can then be integrated and used in various natural language processing (NLP) applications, such as textual entailment, summarization, question answering, and explicit reasoning. Firstly, we surveyed neural OpenIE. Thanks to the rapid development of deep learning technologies, numerous neural OpenIE architectures have been proposed, achieving considerable performance improvement. In this survey, we provide an extensive overview of the state-of-the-art neural OpenIE models, their key design decisions, strengths, and weaknesses. Then, we discuss the limitations of current solutions and the open issues in the OpenIE problem itself. Finally, we list recent trends that could help expand its scope and applicability, setting up promising directions for future research in OpenIE. This paper is the first review of neural OpenIE. Secondly, we consider semi-supervised generative OpenIE to boost the performance by utilizing unlabeled texts. Deep learning-based OpenIE systems need more high-quality training data. Moreover, as mentioned in previous work, different OpenIE applications focus on different types of information. Existing unsupervised data bootstrapping solutions need help producing high-quality training data and being more flexible in adjusting the data to be suitable for specific applications. Recently, generative methods have shown great potential in various NLP tasks. We leverage a mean teacher-based semi-supervised learning method to boost the performance of a generative OpenIE model. We first pre-train the model using existing training data and then conduct semi-supervised training using gold data and unlabeled corpus. To filter out non-trustworthy training examples generated during the semi-supervised learning phase, we introduce a verification mechanism to improve the quality of the examples. Instead of training two separate models, we use prompts to train a joint model with extraction and verification capabilities. We conduct experiments on OpenIE benchmarks and show the promising results achieved by our proposed method. Finally, we conduct a few experiments evaluating ChatGPT’s zero-shot capability, discuss possible future work directions, and conclude the report.
DOI: 10.32657/10356/171930
Schools: School of Computer Science and Engineering 
Rights: This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0).
Fulltext Permission: open
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Theses

Files in This Item:
File Description SizeFormat 
MasterThesis_ZhouShaowen.pdfMaster thesis967.5 kBAdobe PDFThumbnail

Page view(s)

Updated on Jul 18, 2024

Download(s) 50

Updated on Jul 18, 2024

Google ScholarTM




Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.