Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/184549
Title: | Reinforcement learning with possibility theory | Authors: | Tejas Gupta | Keywords: | Mathematical Sciences | Issue Date: | 2025 | Publisher: | Nanyang Technological University | Source: | Tejas Gupta (2025). Reinforcement learning with possibility theory. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184549 | Abstract: | Deep Reinforcement Learning (RL) agents often struggle to balance exploration and exploitation due to their value estimates not accounting for epistemic uncertainty. Possibility theory, with its maxitive calculus and less restrictive normalization constraints, offers a method for tracking and using epistemic uncertainty. This thesis investigates whether possibilistic modeling of uncertainty can drive principled optimistic exploration. Three algorithmic variants are proposed: (i) Possibilistic DQN, which models Q-values via a Gaussian and acts via a parameter-free closed-form maximum expected value; (ii) Possibilistic Q-Ensembles, which maintain possibility weighting over multiple Q-networks and update them using a possibilistic Bayes rule; and (iii) Possibilistic Model-Based Learning, a pair of zero- and one-step planning algorithms that propagate optimistic targets from possibilistic models. The models are tested on benchmark environments along with sparse and stochastic variants. The study confirms the efficacy of maximum-expected-value optimism in the first method, particularly in deterministic and sparse environments. The ensemble possibilistic approach outperforms standard non-weighted ensembles and DQN. The third model underperformed relative to baselines; further work is needed to tune hyperparameters and improve model accuracy. Overall, we demonstrate the utility of possibility theory to capture uncertainty and make exploration more efficient. | URI: | https://hdl.handle.net/10356/184549 | Schools: | School of Physical and Mathematical Sciences | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | SPMS Student Reports (FYP/IA/PA/PI) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Reinforcement Learning with Possibility Theory.pdf Restricted Access | 4.47 MB | Adobe PDF | View/Open |
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.