A discrete-time POMDP models the relationship between an agent and its environment. Formally, a POMDP is a 7-tuple (S,A,T,R,\Omega ,O,\gamma ), where
S is a set of states,
A is a set of actions,
T is a set of conditional transition probabilities between states,
R: S \times A \to \mathbb{R} is the reward function.
\Omega is a set of observations,
O is a set of conditional observation probabilities, and
\gamma \in [0, 1] is the discount factor.
It seems like POMDPs are an established framework for modeling agents that interact with an environment. Can they be explained with this CT machinery?
Hi Jaz,
How do your open markov systems relate to things like Markov Decision Processes and Partially Observed Markov Decision Processes.
From wikipedia:
It seems like POMDPs are an established framework for modeling agents that interact with an environment. Can they be explained with this CT machinery?