Skip to content

pomdp_1.0.2

Compare
Choose a tag to compare
@mhahsler mhahsler released this 17 May 20:09
· 154 commits to master since this release
  • policy_graph() can now produce policy trees for finite-horizon problems, and the initial belief can be specified.
  • simulate_POMDP(): fixed bug with not using horizon.
  • reward() and reward_node_action() have now been separated.
  • sample_belief_space() gained method 'trajectories'.
  • simulate_POMDP(): supports not epsilon-greedy policies.
  • added x_prob() and x_val() functions to access individual parts of the matrices.
  • fixed converged finite-horizon case. It now only returns the converged graph/alpha.
  • we use not internally NA to represent * in the POMDP definition.
  • actions, states, and observations are now factors in most places.