Please use this identifier to cite or link to this item:
Title: Universal adversarial network attacks on traffic light recognition of Apollo autonomous driving system
Authors: Chia, Yi You
Keywords: Engineering::Computer science and engineering
Issue Date: 2022
Publisher: Nanyang Technological University
Source: Chia, Y. Y. (2022). Universal adversarial network attacks on traffic light recognition of Apollo autonomous driving system. Final Year Project (FYP), Nanyang Technological University, Singapore.
Abstract: Autonomous Vehicles are becoming increasingly important and relevant in today’s world. Their applications can be found everywhere, from public transport to overcome land and workforce constraints to personal uses for convenience to business uses for freight transportation and utility services sectors. Therefore, emphasising the importance of safety in these autonomous vehicles. Autonomous vehicles use Autonomous Driving Systems (ADS), which requires inputs from multiple camera sensors to be passed into a machine learning model to output the results that directly control the car movements. This paper focuses on the safety of these machine learning models. A black-box Universal Adversarial Network (UAN) is first trained to create a universal perturbation, which will be used to attack the machine learning model that recognises traffic light signals. Eventually producing a wrong traffic signal as an output. Multiple variations of the UAN are produced to study their effect on the accuracy of these machine learning models. This vulnerability will also be studied in a realistic environment using Baidu Apollo ADS and LGSVL. Lastly, basic defences of Apollo ADS will be explored.
Schools: School of Computer Science and Engineering 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
Chia Yi You_FYP_Final Report.pdf
  Restricted Access
2.11 MBAdobe PDFView/Open

Page view(s)

Updated on Sep 30, 2023

Download(s) 50

Updated on Sep 30, 2023

Google ScholarTM


Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.