POMDP Usage
POMDP
The MDP struct gives the following:
γ: discount factor𝒮: state space𝒜: action space𝒪: observation spaceT: transition functionR: reward functionO: observation functionTRO: function that allows us to sample transition, reward, and observation
The function T takes in a state s and an action a and returns a distribution of possible states. The reward function R takes in a state s and action a and returns an reward. The observation function O takes in a state s and an action a and returns a distribution of possible observations. Finally TRO takes in a state s and an action a and returns a tuple (s', r, o) where s' is the new state sampled from the transition function, r is the reward and o is an observation sampled from the observation function.