data("Tiger")
# Infinite horizon
sol <- solve_POMDP(model = Tiger)
sol
# policy with value function, optimal action and transitions for observations.
policy(sol)
plot_value_function(sol)
# Finite horizon (we use incremental pruning because grid does not converge)
sol <- solve_POMDP(model = Tiger, method = "incprune",
horizon = 3, discount = 1)
sol
policy(sol)
# Note: We see that it is initially better to listen till we make
# a decision in the final epoch.
# MDP policy
data(Maze)
sol <- solve_MDP(Maze)
policy(sol)
Run the code above in your browser using DataLab