Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/184549
Title: Reinforcement learning with possibility theory
Authors: Tejas Gupta
Keywords: Mathematical Sciences
Issue Date: 2025
Publisher: Nanyang Technological University
Source: Tejas Gupta (2025). Reinforcement learning with possibility theory. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184549
Abstract: Deep Reinforcement Learning (RL) agents often struggle to balance exploration and exploitation due to their value estimates not accounting for epistemic uncertainty. Possibility theory, with its maxitive calculus and less restrictive normalization constraints, offers a method for tracking and using epistemic uncertainty. This thesis investigates whether possibilistic modeling of uncertainty can drive principled optimistic exploration. Three algorithmic variants are proposed: (i) Possibilistic DQN, which models Q-values via a Gaussian and acts via a parameter-free closed-form maximum expected value; (ii) Possibilistic Q-Ensembles, which maintain possibility weighting over multiple Q-networks and update them using a possibilistic Bayes rule; and (iii) Possibilistic Model-Based Learning, a pair of zero- and one-step planning algorithms that propagate optimistic targets from possibilistic models. The models are tested on benchmark environments along with sparse and stochastic variants. The study confirms the efficacy of maximum-expected-value optimism in the first method, particularly in deterministic and sparse environments. The ensemble possibilistic approach outperforms standard non-weighted ensembles and DQN. The third model underperformed relative to baselines; further work is needed to tune hyperparameters and improve model accuracy. Overall, we demonstrate the utility of possibility theory to capture uncertainty and make exploration more efficient.
URI: https://hdl.handle.net/10356/184549
Schools: School of Physical and Mathematical Sciences 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SPMS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
Reinforcement Learning with Possibility Theory.pdf
  Restricted Access
4.47 MBAdobe PDFView/Open

Page view(s)

12
Updated on May 7, 2025

Download(s)

2
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.