View source: R/optimal_action.R
optimal_action | R Documentation |
Determines the optimal action for a policy (solved POMDP) for a given belief at a given epoch.
optimal_action(model, belief = NULL, epoch = 1)
model |
a solved POMDP. |
belief |
The belief (probability distribution over the states) as a
vector or a matrix with multiple belief states as rows. If |
epoch |
what epoch of the policy should be used. Use 1 for converged policies. |
The name of the optimal action.
Michael Hahsler
Other policy:
estimate_belief_for_nodes()
,
plot_belief_space()
,
plot_policy_graph()
,
policy()
,
policy_graph()
,
projection()
,
reward()
,
solve_POMDP()
,
solve_SARSOP()
,
value_function()
data("Tiger")
Tiger
sol <- solve_POMDP(model = Tiger)
# these are the states
sol$states
# belief that tiger is to the left
optimal_action(sol, c(1, 0))
optimal_action(sol, "tiger-left")
# belief that tiger is to the right
optimal_action(sol, c(0, 1))
optimal_action(sol, "tiger-right")
# belief is 50/50
optimal_action(sol, c(.5, .5))
optimal_action(sol, "uniform")
# the POMDP is converged, so all epoch give the same result.
optimal_action(sol, "tiger-right", epoch = 10)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.