Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/181155
Title: Haze removal from an image or a video via generative adversarial networks
Authors: Chen, Zhong Jiang
Keywords: Computer and Information Science
Issue Date: 2024
Publisher: Nanyang Technological University
Source: Chen, Z. J. (2024). Haze removal from an image or a video via generative adversarial networks. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181155
Project: SCSE22-0577
Abstract: Low visibility caused by haze and fog is one of the major reasons for traffic and aviation accidents. This paper introduces a more easy-to-access solution to remove haze from a single image, video, and live-streaming. My approach uses a modified conditional Generative Adversarial Network (cGAN) with a DenseNet-121 architecture to efficiently dehaze visual inputs. Unlike models that use Tiramisu [5] or depend on two-step pipelines, The modified model ensures the accuracy of structure and clarity of the visual by removing haze by optimizing the generator-discriminator interaction within the GAN framework. The effectiveness of the modified model is demonstrated through a comprehensive experiment on synthetic and real-world data, obtaining competitive results in PSNR, SSIM, and subjective quality measures. This system aims to improve visibility in live-streaming scenarios, such as for vehicles and aircraft, potentially reducing the probability of accidents under low-visibility conditions.
URI: https://hdl.handle.net/10356/181155
Schools: College of Computing and Data Science 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
SCSE22-0577_Chen ZhongJiang.pdf
  Restricted Access
757.03 kBAdobe PDFView/Open

Page view(s)

76
Updated on May 7, 2025

Download(s)

12
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.