Infrastructure for Partially Observable Markov Decision Processes (POMDP)


[Up] [Top]

Documentation for package ‘pomdp’ version 1.0.1

Help Pages

pomdp-package pomdp: Infrastructure for Partially Observable Markov Decision Processes (POMDP)
approx_MDP_policy_evaluation Solve an MDP Problem
estimate_belief_for_nodes Visualize a POMDP Policy Graph
Maze Steward Russell's 4x3 Maze MDP
maze Steward Russell's 4x3 Maze MDP
MDP Define an MDP Problem
MDP2POMDP Define an MDP Problem
observation_matrix Extract the Transition, Observation or Reward Information from a POMDP
optimal_action Optimal action for a belief
O_ Define a POMDP Problem
plot_belief_space Plot a 2D or 3D Projection of the Belief Space
plot_policy_graph Visualize a POMDP Policy Graph
plot_value_function Plot the Value Function of a POMDP Solution
policy Extract the Policy from a POMDP/MDP
policy_graph Visualize a POMDP Policy Graph
POMDP Define a POMDP Problem
q_values_MDP Solve an MDP Problem
random_MDP_policy Solve an MDP Problem
read_POMDP Read and write a POMDP Model to a File in POMDP Format
reward Calculate the Reward for a POMDP Solution
reward_matrix Extract the Transition, Observation or Reward Information from a POMDP
round_stochastic Round a stochastic vector or a row-stochastic matrix
R_ Define a POMDP Problem
sample_belief_space Sample from the Belief Space
simulate_POMDP Simulate Trajectories in a POMDP
solve_MDP Solve an MDP Problem
solve_POMDP Solve a POMDP Problem using pomdp-solver
solve_POMDP_parameter Solve a POMDP Problem using pomdp-solver
solve_SARSOP Solve a POMDP Problem using SARSOP
Three_doors Tiger Problem POMDP Specification
Tiger Tiger Problem POMDP Specification
transition_matrix Extract the Transition, Observation or Reward Information from a POMDP
T_ Define a POMDP Problem
update_belief Belief Update
write_POMDP Read and write a POMDP Model to a File in POMDP Format