| pomdp-package | pomdp: Infrastructure for Partially Observable Markov Decision Processes (POMDP) |
| approx_MDP_policy_evaluation | Solve an MDP Problem |
| estimate_belief_for_nodes | Visualize a POMDP Policy Graph |
| Maze | Steward Russell's 4x3 Maze MDP |
| maze | Steward Russell's 4x3 Maze MDP |
| MDP | Define an MDP Problem |
| MDP2POMDP | Define an MDP Problem |
| observation_matrix | Extract the Transition, Observation or Reward Information from a POMDP |
| optimal_action | Optimal action for a belief |
| O_ | Define a POMDP Problem |
| plot_belief_space | Plot a 2D or 3D Projection of the Belief Space |
| plot_policy_graph | Visualize a POMDP Policy Graph |
| plot_value_function | Plot the Value Function of a POMDP Solution |
| policy | Extract the Policy from a POMDP/MDP |
| policy_graph | Visualize a POMDP Policy Graph |
| POMDP | Define a POMDP Problem |
| q_values_MDP | Solve an MDP Problem |
| random_MDP_policy | Solve an MDP Problem |
| read_POMDP | Read and write a POMDP Model to a File in POMDP Format |
| reward | Calculate the Reward for a POMDP Solution |
| reward_matrix | Extract the Transition, Observation or Reward Information from a POMDP |
| round_stochastic | Round a stochastic vector or a row-stochastic matrix |
| R_ | Define a POMDP Problem |
| sample_belief_space | Sample from the Belief Space |
| simulate_POMDP | Simulate Trajectories in a POMDP |
| solve_MDP | Solve an MDP Problem |
| solve_POMDP | Solve a POMDP Problem using pomdp-solver |
| solve_POMDP_parameter | Solve a POMDP Problem using pomdp-solver |
| solve_SARSOP | Solve a POMDP Problem using SARSOP |
| Three_doors | Tiger Problem POMDP Specification |
| Tiger | Tiger Problem POMDP Specification |
| transition_matrix | Extract the Transition, Observation or Reward Information from a POMDP |
| T_ | Define a POMDP Problem |
| update_belief | Belief Update |
| write_POMDP | Read and write a POMDP Model to a File in POMDP Format |