POMDPs.jl
A Julia interface for defining, solving and simulating partially observable Markov decision processes and their fully observable counterparts.
Package and Ecosystem Features
- General interface that can handle problems with discrete and continuous state/action/observation spaces
- A number of popular state-of-the-art solvers implemented for use out-of-the-box
- Tools that make it easy to define problems and simulate solutions
- Simple integration of custom solvers into the existing interface
Available Packages
The POMDPs.jl package contains only the interface used for expressing and solving Markov decision processes (MDPs) and partially observable Markov decision processes (POMDPs). The POMDPTools package acts as a "standard library" for the POMDPs.jl interface, providing implementations of commonly-used components such as policies, belief updaters, distributions, and simulators. The list of solver and support packages maintained by the JuliaPOMDP community is available at the POMDPs.jl Readme.
Documentation Outline
Documentation comes in three forms:
- An explanatory guide is available in the sections outlined below.
- How-to examples are available throughout this documentation with specicic examples in Examples and Gallery of POMDPs.jl Problems.
- Reference docstrings for the entire POMDPs.jl interface are available in the API Documentation section.
When updating these documents, make sure this is synced with docs/make.jl!!
Basics
Defining POMDP Models
- Defining POMDPs and MDPs
- Spaces and Distributions
Writing Solvers and Updaters
- Solvers
- Example: Defining an offline solver
- Example: Defining an online solver
- Defining a Belief Updater
Analyzing Results
Examples and Gallery
- Examples
- Defining a POMDP
- Using Different Solvers
- Simulations Examples
- GridWorld MDP Tutorial
- Dependencies
- Problem Overview
- Defining the Grid World MDP Type
- Grid World State Space
- Grid World Action Space
- Grid World Transition Function
- Grid World Reward Function
- Grid World Remaining Functions
- Solving the Grid World MDP (Value Iteration)
- Solving the Grid World MDP (MCTS)
- Visualizing the Value Iteration Policy
- Seeing a Policy In Action
- Gallery of POMDPs.jl Problems
POMDPTools - the standard library for POMDPs.jl
- POMDPTools: the standard library for POMDPs.jl
- Implemented Distributions
- Model Tools
- Visualization
- Implemented Belief Updaters
- Implemented Policies
- Policy Evaluation
- Implemented Simulators
- CommonRLInterface Integration
- Testing
Reference
- Frequently Asked Questions (FAQ)
- What is the difference between
transition
,gen
, and@gen
? - How do I save my policies?
- Why is my solver producing a suboptimal policy?
- What if I don't use the
rng
argument? - Why are all the solvers in separate modules?
- How can I implement terminal actions?
- Why are there two versions of
reward
? - How do I implement
reward(m, s, a)
if the reward depends on the next state? - Why do I need to put type assertions pomdp::POMDP into the function signature?
- What is the difference between
- API Documentation