Skip to content

Latest commit

 

History

History
25 lines (16 loc) · 1.28 KB

README.md

File metadata and controls

25 lines (16 loc) · 1.28 KB

irl-es

  • inverse reinforcement learning for control through evolutionary strategies

Demonstration Trajectory Generation

For generating demonstration for toy control problems through trajectory optimization, we use pydrake. This custom colab notebook comprises of the i)setup ii) trajectoy generation iii) Visualisation and iv) other boiler plates. At present it is ready for pendulumn, and one can easily download a zipped folder of the trajectory for a given set of dynamics and simulation config.

Guided Cost Learning

This repo is a working implementation of guided cost learning (https://arxiv.org/abs/1603.00448) for discrete action space. Currently, it supports the CartPole-v0 environment.

How to run:

$ python3.8 main.py 

Description of files:

  • generating_expert.py: Generates an expert on CartPole, by training vanilla policy gradient, and finally stores trained trajecteries as expert samples at expert_samples.
  • experts/PG.py: Implementation of vanilla policy gradient. This is reused at several places.
  • main.py: Contains the main implementation of GCL.

Results: