Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/180288
Full metadata record
DC FieldValueLanguage
dc.contributor.authorLow, Ashton Kin Yunen_US
dc.contributor.authorNimrod, Lilithen_US
dc.contributor.authorAlam, Sameeren_US
dc.contributor.authorPoh, Leston Choo Kiaten_US
dc.date.accessioned2024-10-01T08:43:44Z-
dc.date.available2024-10-01T08:43:44Z-
dc.date.issued2024-
dc.identifier.citationLow, A. K. Y., Nimrod, L., Alam, S. & Poh, L. C. K. (2024). Deep neural network-based automatic speech recognition for ATC-pilot audio transcription. 2024 International Conference on Research in Air Transportation (ICRAT).en_US
dc.identifier.urihttps://hdl.handle.net/10356/180288-
dc.description.abstractArtificial Intelligence (AI) has demonstrated ability to manage complex processes highly effectively, and thus is widely seen as a key component in future airport ATM systems. Future AI tools for ATM will rely on digital data, such as surveillance, radar, weather, flight plans, for their operation. However, the foundational Air Traffic Control Officer (ATCo)-pilot communication medium is voice, which is a vital source of situational data. Controller Pilot Data Link Communications (CPDLC) has been developed as an alternative, text-based communication delivery method, however ATCo-pilot communications will not be completed transitioned to this framework in the near-term future. Moreover, as CPDLC is a one-to-one communication paradigm, the additional situational awareness of other traffic provided by traditional party-line VHF communications is potentially lost. Therefore, an automated speech to-text translation tool can be seen as a missing link, enabling traditional ATCo-pilot voice communications to be automatically translated and input into a datalink system such as CPDLC. To this end this paper presents a Machine Learning (ML) based Automatic Speech Recognition (ASR) framework that is able to accurately translate ATCo-pilot speech communication to text, achieving a Word Error Rate of only 6.13%. Moreover, the presented model is able to extract seven entities with an accuracy and F1-score of 91.8% and 84.4% respectively, which is similar to previously presented models but can only capable of extracting three. A detailed design of the framework is provided to enable its replication by the wider research community.en_US
dc.language.isoenen_US
dc.relationNTU Ref: 2017-1619en_US
dc.rights© 2024 ICRAT. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at https://www.icrat.org/upcoming-conference/papers/.en_US
dc.subjectEngineeringen_US
dc.titleDeep neural network-based automatic speech recognition for ATC-pilot audio transcriptionen_US
dc.typeConference Paperen
dc.contributor.schoolSchool of Mechanical and Aerospace Engineeringen_US
dc.contributor.conference2024 International Conference on Research in Air Transportation (ICRAT)en_US
dc.contributor.researchAir Traffic Management Research Instituteen_US
dc.description.versionPublished versionen_US
dc.identifier.urlhttps://www.icrat.org/upcoming-conference/papers/-
dc.identifier.urlhttps://www.icrat.org/-
dc.subject.keywordsArtificial intelligenceen_US
dc.subject.keywordsMachine learningen_US
dc.subject.keywordsAutomatic speech recognitionen_US
dc.subject.keywordsAir traffic controlen_US
dc.subject.keywordsATCo-pilot communicationen_US
dc.citation.conferencelocationSingaporeen_US
dc.description.acknowledgementThis research is supported by Saab Sweden under its Machine Learning For Airport Management And Tower Control project at NTU Singapore.en_US
item.grantfulltextopen-
item.fulltextWith Fulltext-
Appears in Collections:MAE Conference Papers
Files in This Item:
File Description SizeFormat 
ICRAT2024_paper_88.pdf1.17 MBAdobe PDFThumbnail
View/Open

Page view(s)

29
Updated on Oct 9, 2024

Download(s)

2
Updated on Oct 9, 2024

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.