Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/156966
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Leung, Jonathan | en_US |
dc.contributor.author | Shen, Zhiqi | en_US |
dc.contributor.author | Zeng, Zhiwei | en_US |
dc.contributor.author | Miao, Chunyan | en_US |
dc.date.accessioned | 2022-05-12T00:48:28Z | - |
dc.date.available | 2022-05-12T00:48:28Z | - |
dc.date.issued | 2021 | - |
dc.identifier.citation | Leung, J., Shen, Z., Zeng, Z. & Miao, C. (2021). Goal modelling for deep reinforcement learning agents. Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD 2021), 12975, 271-286. https://dx.doi.org/10.1007/978-3-030-86486-6_17 | en_US |
dc.identifier.isbn | 9783030864859 | - |
dc.identifier.issn | 0302-9743 | - |
dc.identifier.uri | https://hdl.handle.net/10356/156966 | - |
dc.description.abstract | Goals provide a high-level abstraction of an agent’s objectives and guide its behavior in complex environments. As agents become more intelligent, it is necessary to ensure that the agent’s goals are aligned with the goals of the agent designers to avoid unexpected or unwanted agent behavior. In this work, we propose using Goal Net, a goal-oriented agent modelling methodology, as a way for agent designers to incorporate their prior knowledge regarding the subgoals an agent needs to achieve in order to accomplish an overall goal. This knowledge is used to guide the agent’s learning process to train it to achieve goals in dynamic environments where its goal may change between episodes. We propose a model that integrates a Goal Net model and hierarchical reinforcement learning. A high-level goal selection policy selects goals according to a given Goal Net model and a low-level action selection policy selects actions based on the selected goal, both of which use deep neural networks to enable learning in complex, high-dimensional environments. The experiments demonstrate that our method is more sample efficient and can obtain higher average rewards than other related methods that incorporate prior human knowledge in similar ways. | en_US |
dc.description.sponsorship | National Research Foundation (NRF) | en_US |
dc.language.iso | en | en_US |
dc.relation | NRF-NRFI05-2019-0002 | en_US |
dc.rights | © 2021 Springer Nature Switzerland AG. All rights reserved. This paper was published in Proceedings of Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD 2021) and is made available with permission of Springer Nature Switzerland AG. | en_US |
dc.subject | Engineering::Computer science and engineering | en_US |
dc.title | Goal modelling for deep reinforcement learning agents | en_US |
dc.type | Conference Paper | en |
dc.contributor.school | School of Computer Science and Engineering | en_US |
dc.contributor.conference | Joint European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD 2021) | en_US |
dc.identifier.doi | 10.1007/978-3-030-86486-6_17 | - |
dc.description.version | Submitted/Accepted version | en_US |
dc.identifier.scopus | 2-s2.0-85115448159 | - |
dc.identifier.volume | 12975 | en_US |
dc.identifier.spage | 271 | en_US |
dc.identifier.epage | 286 | en_US |
dc.subject.keywords | Deep Reinforcement Learning | en_US |
dc.subject.keywords | Hierarchical Reinforcement Learning | en_US |
dc.citation.conferencelocation | Bilbao, Spain | en_US |
dc.description.acknowledgement | This research is supported, in part, by the National Research Foundation, Prime Minister’s Office, Singapore under its NRF Investigatorship Programme (NRF Award No. NRF-NRFI05-2019-0002). | en_US |
item.grantfulltext | embargo_20220917 | - |
item.fulltext | With Fulltext | - |
Appears in Collections: | SCSE Conference Papers |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Goal Modelling for Deep Reinforcement Learning Agents.pdf Until 2022-09-17 | 2.15 MB | Adobe PDF | Under embargo until Sep 17, 2022 |
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.