This repo contains code for scalable IRL approach that involves two key steps:
- Representation learning via adaptive state graphs aka controller graphs
- IRL over (infinite) state and action spaces via sampled trajectories
- Efficient and flexible graph based (hierarchical) representation.
- Ability to incoporate task specific constrains directly into the MDP representation (the graph).
- Admits efficient IRL algorithms on sampled trajectories.
- Currently on BIRL variants are implemented
git clone https://github.com/makokal/scalable-irl.git
cd scalable-irl
[sudo] pip install -r requirements.txt # install dependencies
make
[sudo] make develop # For local development without global install
[sudo] make install # for global install
See examples folder.
- More value approximation/projection methods (e.g. Nystrom)
- More guided sampling strategies/heuristics
- Model-free RL solvers
- Additional IRL variants, e.g. LP, MaxEnt
Pull requests, issues are always welcome
If you use this software in your work, please cite the following paper
@InProceedings{okalRSSLfd15,
author = {Okal, Billy and Gilbert, Hugo and Arras, Kai O.},
title = {Efficient Inverse Reinforcement Learning using Adaptive State Graphs},
booktitle={Robotics: Science and Systems (RSS), Workshop on Learning from Demonstration: Inverse optimal control, Reinforcement learning and Lifelong learning},
address = {Rome, Italy},
year={2015},
}