Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/146557
Full metadata record
DC FieldValueLanguage
dc.contributor.authorDong, Linsenen_US
dc.date.accessioned2021-03-01T05:44:21Z-
dc.date.available2021-03-01T05:44:21Z-
dc.date.issued2021-
dc.identifier.citationDong, L. (2021). Baconian : a unified model-based reinforcement learning library. Master's thesis, Nanyang Technological University, Singapore.en_US
dc.identifier.urihttps://hdl.handle.net/10356/146557-
dc.description.abstractReinforcement Learning (RL) has become a trending research topic with great success in outperforming humans on many tasks including video games, board games, and robotics control. By leveraging Deep Learning (DL), RL algorithms can consume a large volume of data without any prior knowledge of the system dynamics. However, requiring a large amount of data also limits the applicability in many fields where data is costly to obtain. Model-based Reinforcement Learning (MBRL) is regarded as a promising way to achieve high data efficiency while maintaining comparable performance. MBRL equips a dynamic transition model to facilitate and speed up the policy searching by learning the system dynamics. But there are no satisfying open-sourced libraries for the RL community to conduct MBRL research. Therefore, to fill the gap, we propose an open-sourced, flexible, and user-friendly MBRL library, Baconian, to facilitate the research on MBRL. In this thesis, we illustrate the library from the aspects of design principle, implementations, and the programming guide. Various benchmark results are also given. To reach high flexibility, modularized design is applied by separating the library into three components: Experiment Manager, Training Engine, and Monitor. For implementations, we provide commonly used functionalities including parameter management, TensorFlow integration etc. Moreover, we utilize Baconian to conduct RL experiments in real research topics at the case study section. First, we utilize Baconian as the framework to tune the Dyna-style MBRL hyper-parameters in an online fashion. Our proposed method reaches a similar or better performance out of all five tasks compared to three baseline methods. Second, we use Baconian to apply RL algorithms for online video bitrate selection optimization where our method outperforms the best baseline method on average bitrate metric by 7.8%.en_US
dc.language.isoenen_US
dc.publisherNanyang Technological Universityen_US
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0).en_US
dc.subjectEngineering::Computer science and engineering::Computing methodologies::Artificial intelligenceen_US
dc.titleBaconian : a unified model-based reinforcement learning libraryen_US
dc.typeThesis-Master by Researchen_US
dc.contributor.supervisorWen Yonggangen_US
dc.contributor.schoolSchool of Computer Science and Engineeringen_US
dc.description.degreeMaster of Engineeringen_US
dc.contributor.researchCloud Application and Platform Laben_US
dc.identifier.doi10.32657/10356/146557-
dc.contributor.supervisoremailYGWEN@ntu.edu.sgen_US
item.grantfulltextopen-
item.fulltextWith Fulltext-
Appears in Collections:SCSE Theses
Files in This Item:
File Description SizeFormat 
DongLinsen-MENG-SCSE-NTU-Thesis-revised-v1-20210206.pdf2.93 MBAdobe PDFThumbnail
View/Open

Page view(s)

477
Updated on Apr 26, 2025

Download(s) 20

303
Updated on Apr 26, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.