Simulations Examples
In these simulation examples, we will use the crying baby POMDPs defined in the Defining a POMDP section (i.e. quick_crying_baby_pomdp
, explicit_crying_baby_pomdp
, gen_crying_baby_pomdp
, and tabular_crying_baby_pomdp
).
Stepthrough
The stepthrough simulater provides a window into the simulation with a for-loop syntax.
Within the body of the for loop, we have access to the belief, the action, the observation, and the reward, in each step. We also calculate the sum of the rewards in this example, but note that this is not the discounted reward.
policy = RandomPolicy(quick_crying_baby_pomdp)
r_sum = 0.0
step = 0
for (b, s, a, o, r) in stepthrough(quick_crying_baby_pomdp, policy, DiscreteUpdater(quick_crying_baby_pomdp), "b,s,a,o,r"; max_steps=4)
step += 1
println("Step $step")
println("b = sated => $(b.b[1]), hungry => $(b.b[2])")
@show s
@show a
@show o
@show r
r_sum += r
@show r_sum
println()
end
Step 1
b = sated => 1.0, hungry => 0.0
s = :sated
a = :ignore
o = :crying
r = 0.0
r_sum = 0.0
Step 2
b = sated => 0.5294117647058822, hungry => 0.47058823529411764
s = :hungry
a = :sing
o = :quiet
r = -10.5
r_sum = -10.5
Step 3
b = sated => 0.9010011123470524, hungry => 0.09899888765294775
s = :hungry
a = :feed
o = :quiet
r = -15.0
r_sum = -25.5
Step 4
b = sated => 1.0, hungry => 0.0
s = :sated
a = :ignore
o = :quiet
r = 0.0
r_sum = -25.5
Rollout Simulations
While stepthrough is a flexible and convenient tool for many user-facing demonstrations, it is often less error-prone to use the standard simulate function with a Simulator
object. The simplest Simulator is the RolloutSimulator
. It simply runs a simulation and returns the discounted reward.
policy = RandomPolicy(explicit_crying_baby_pomdp)
sim = RolloutSimulator(max_steps=10)
r_sum = simulate(sim, explicit_crying_baby_pomdp, policy)
println("Total discounted reward: $r_sum")
Total discounted reward: -40.8528906845
Recording Histories
Sometimes it is important to record the entire history of a simulation for further examination. This can be accomplished with a HistoryRecorder
.
policy = RandomPolicy(tabular_crying_baby_pomdp)
hr = HistoryRecorder(max_steps=5)
history = simulate(hr, tabular_crying_baby_pomdp, policy, DiscreteUpdater(tabular_crying_baby_pomdp), Deterministic(1))
The history object produced by a HistoryRecorder
is a SimHistory
, documented in the POMDPTools simulater section Histories. The information in this object can be accessed in several ways. For example, there is a function:
discounted_reward(history)
-6.097550000000001
Accessor functions like state_hist
and action_hist
can also be used to access parts of the history:
state_hist(history)
6-element Vector{Int64}:
1
1
1
1
1
1
collect(action_hist(history))
5-element Vector{Int64}:
2
1
2
2
2
Keeping track of which states, actions, and observations belong together can be tricky (for example, since there is a starting state, and ending state, but no action is taken from the ending state, the list of actions has a different length than the list of states). It is often better to think of histories in terms of steps that include both starting and ending states.
The most powerful function for accessing the information in a SimHistory
is the eachstep
function which returns an iterator through named tuples representing each step in the history. The eachstep
function is similar to the stepthrough
function above except that it iterates through the immutable steps of a previously simulated history instead of conducting the simulation as the for loop is being carried out.
r_sum = 0.0
step = 0
for step_i in eachstep(sim_history, "b,s,a,o,r")
step += 1
println("Step $step")
println("step_i.b = sated => $(step_i.b.b[1]), hungry => $(step_i.b.b[2])")
@show step_i.s
@show step_i.a
@show step_i.o
@show step_i.r
r_sum += step_i.r
@show r_sum
println()
end
end # hide
Step 1
step_i.b = sated => 1.0, hungry => 0.0
step_i.s = 1
step_i.a = 2
step_i.o = 2
step_i.r = -0.5
r_sum = -0.5
Step 2
step_i.b = sated => 0.989010989010989, hungry => 0.010989010989010992
step_i.s = 1
step_i.a = 1
step_i.o = 2
step_i.r = -5.0
r_sum = -5.5
Step 3
step_i.b = sated => 1.0, hungry => 0.0
step_i.s = 1
step_i.a = 2
step_i.o = 2
step_i.r = -0.5
r_sum = -6.0
Step 4
step_i.b = sated => 0.989010989010989, hungry => 0.010989010989010992
step_i.s = 1
step_i.a = 2
step_i.o = 2
step_i.r = -0.5
r_sum = -6.5
Step 5
step_i.b = sated => 0.9878048780487805, hungry => 0.012195121951219514
step_i.s = 1
step_i.a = 2
step_i.o = 2
step_i.r = -0.5
r_sum = -7.0
Parallel Simulations
It is often useful to evaluate a policy by running many simulations. The parallel simulator is the most effective tool for this. To use the parallel simulator, first create a list of Sim
objects, each of which contains all of the information needed to run a simulation. Then then run the simulations using run_parallel
, which will return a DataFrame
with the results.
In this example, we will compare the performance of the polcies we computed in the Using Different Solvers section (i.e. sarsop_policy
, pomcp_planner
, and heuristic_policy
). To evaluate the policies, we will run 100 simulations for each policy. We can do this by adding 100 Sim
objects of each policy to the list.
using DataFrames
using StatsBase: std
# Defining paramters for the simulations
number_of_sim_to_run = 100
max_steps = 20
starting_seed = 1
# We will also compare against a random policy
rand_policy = RandomPolicy(quick_crying_baby_pomdp, rng=MersenneTwister(1))
# Create the list of Sim objects
sim_list = []
# Add 100 Sim objects of each policy to the list.
for sim_number in 1:number_of_sim_to_run
seed = starting_seed + sim_number
# Add the SARSOP policy
push!(sim_list, Sim(
quick_crying_baby_pomdp,
rng=MersenneTwister(seed),
sarsop_policy,
max_steps=max_steps,
metadata=Dict(:policy => "sarsop", :seed => seed))
)
# Add the POMCP policy
push!(sim_list, Sim(
quick_crying_baby_pomdp,
rng=MersenneTwister(seed),
pomcp_planner,
max_steps=max_steps,
metadata=Dict(:policy => "pomcp", :seed => seed))
)
# Add the heuristic policy
push!(sim_list, Sim(
quick_crying_baby_pomdp,
rng=MersenneTwister(seed),
heuristic_policy,
max_steps=max_steps,
metadata=Dict(:policy => "heuristic", :seed => seed))
)
# Add the random policy
push!(sim_list, Sim(
quick_crying_baby_pomdp,
rng=MersenneTwister(seed),
rand_policy,
max_steps=max_steps,
metadata=Dict(:policy => "random", :seed => seed))
)
end
# Run the simulations in parallel
data = run_parallel(sim_list)
# Define a function to calculate the mean and confidence interval
function mean_and_ci(x)
m = mean(x)
ci = 1.96 * std(x) / sqrt(length(x)) # 95% confidence interval
return (mean = m, ci = ci)
end
# Calculate the mean and confidence interval for each policy
grouped_df = groupby(data, :policy)
result = combine(grouped_df, :reward => mean_and_ci => AsTable)
Row | policy | mean | ci |
---|---|---|---|
String? | Float64 | Float64 | |
1 | sarsop | -14.6264 | 1.81814 |
2 | pomcp | -18.6904 | 1.57649 |
3 | heuristic | -16.0943 | 1.91474 |
4 | random | -30.4201 | 2.64208 |
By default, the parallel simulator only returns the reward from each simulation, but more information can be gathered by specifying a function to analyze the Sim
-history pair and record additional statistics. Reference the POMDPTools simulator section for more information (Specifying information to be recorded).