MDPtoolbox: Markov Decision Processes Toolbox

The Markov Decision Processes (MDP) toolbox proposes functions related to the resolution of discrete-time Markov Decision Processes: finite horizon, value iteration, policy iteration, linear programming algorithms with some variants and also proposes some functions related to Reinforcement Learning.

AuthorIadine Chades, Guillaume Chapron, Marie-Josee Cros, Frederick Garcia, Regis Sabbadin
Date of publication2017-03-03 18:01:56
MaintainerGuillaume Chapron <gchapron@carnivoreconservation.org>
LicenseBSD_3_clause + file LICENSE
Version4.0.3

View on CRAN

Man pages

mdp_bellman_operator: Applies the Bellman operator

mdp_check: Checks the validity of a MDP

mdp_check_square_stochastic: Checks if a matrix is square and stochastic

mdp_computePpolicyPRpolicy: Computes the transition matrix and the reward matrix for a...

mdp_computePR: Computes a reward matrix for any form of transition and...

mdp_eval_policy_iterative: Evaluates a policy using an iterative method

mdp_eval_policy_matrix: Evaluates a policy using matrix inversion and product

mdp_eval_policy_optimality: Computes sets of 'near optimal' actions for each state

mdp_eval_policy_TD_0: Evaluates a policy using the TD(0) algorithm

mdp_example_forest: Generates a MDP for a simple forest management problem

mdp_example_rand: Generates a random MDP problem

mdp_finite_horizon: Solves finite-horizon MDP using backwards induction algorithm

mdp_LP: Solves discounted MDP using linear programming algorithm

mdp_policy_iteration: Solves discounted MDP using policy iteration algorithm

mdp_policy_iteration_modified: Solves discounted MDP using modified policy iteration...

mdp_Q_learning: Solves discounted MDP using the Q-learning algorithm...

mdp_relative_value_iteration: Solves MDP with average reward using relative value iteration...

mdp_span: Evaluates the span of a vector

MDPtoolbox-package: Markov Decision Processes Toolbox

mdp_value_iteration: Solves discounted MDP using value iteration algorithm

mdp_value_iteration_bound_iter: Computes a bound for the number of iterations for the value...

mdp_value_iterationGS: Solves discounted MDP using Gauss-Seidel's value iteration...

Functions

mdp_bellman_operator Man page
mdp_check Man page
mdp_check_square_stochastic Man page
mdp_computePpolicyPRpolicy Man page
mdp_computePR Man page
mdp_eval_policy_iterative Man page
mdp_eval_policy_matrix Man page
mdp_eval_policy_optimality Man page
mdp_eval_policy_TD_0 Man page
mdp_example_forest Man page
mdp_example_rand Man page
mdp_finite_horizon Man page
mdp_LP Man page
mdp_policy_iteration Man page
mdp_policy_iteration_modified Man page
mdp_Q_learning Man page
mdp_relative_value_iteration Man page
mdp_span Man page
MDPtoolbox Man page
MDPtoolbox-package Man page
mdp_value_iteration Man page
mdp_value_iteration_bound_iter Man page
mdp_value_iterationGS Man page

Files

MDPtoolbox
MDPtoolbox/NAMESPACE
MDPtoolbox/R
MDPtoolbox/R/mdp_policy_iteration.r
MDPtoolbox/R/mdp_value_iteration_bound_iter.r
MDPtoolbox/R/mdp_check.r
MDPtoolbox/R/mdp_computePR.r
MDPtoolbox/R/mdp_eval_policy_matrix.r
MDPtoolbox/R/mdp_LP.r
MDPtoolbox/R/mdp_check_square_stochastic.r
MDPtoolbox/R/mdp_example_rand.r
MDPtoolbox/R/mdp_eval_policy_optimality.r
MDPtoolbox/R/mdp_policy_iteration_modified.r
MDPtoolbox/R/mdp_computePpolicyPRpolicy.r
MDPtoolbox/R/mdp_span.r
MDPtoolbox/R/mdp_relative_value_iteration.r
MDPtoolbox/R/mdp_value_iteration.r
MDPtoolbox/R/mdp_finite_horizon.r
MDPtoolbox/R/mdp_eval_policy_iterative.r
MDPtoolbox/R/mdp_eval_policy_TD_0.r
MDPtoolbox/R/mdp_Q_learning.r
MDPtoolbox/R/mdp_bellman_operator.r
MDPtoolbox/R/mdp_value_iterationGS.r
MDPtoolbox/R/mdp_example_forest.r
MDPtoolbox/MD5
MDPtoolbox/DESCRIPTION
MDPtoolbox/man
MDPtoolbox/man/mdp_span.Rd MDPtoolbox/man/mdp_bellman_operator.Rd MDPtoolbox/man/mdp_eval_policy_matrix.Rd MDPtoolbox/man/mdp_example_forest.Rd MDPtoolbox/man/mdp_value_iteration_bound_iter.Rd MDPtoolbox/man/mdp_value_iterationGS.Rd MDPtoolbox/man/mdp_eval_policy_TD_0.Rd MDPtoolbox/man/mdp_Q_learning.Rd MDPtoolbox/man/mdp_policy_iteration_modified.Rd MDPtoolbox/man/mdp_check.Rd MDPtoolbox/man/mdp_relative_value_iteration.Rd MDPtoolbox/man/MDPtoolbox-package.Rd MDPtoolbox/man/mdp_example_rand.Rd MDPtoolbox/man/mdp_check_square_stochastic.Rd MDPtoolbox/man/mdp_finite_horizon.Rd MDPtoolbox/man/mdp_eval_policy_optimality.Rd MDPtoolbox/man/mdp_computePpolicyPRpolicy.Rd MDPtoolbox/man/mdp_LP.Rd MDPtoolbox/man/mdp_policy_iteration.Rd MDPtoolbox/man/mdp_eval_policy_iterative.Rd MDPtoolbox/man/mdp_computePR.Rd MDPtoolbox/man/mdp_value_iteration.Rd
MDPtoolbox/LICENSE

Questions? Problems? Suggestions? or email at ian@mutexlabs.com.

Please suggest features or report bugs with the GitHub issue tracker.

All documentation is copyright its authors; we didn't write any of that.