ILSwiss is an Easy-to-run Imitation Learning (IL, or Learning from Demonstration, LfD) framework (template) in PyTorch based on existing code base.
The most works are based on rlswiss and rlkit. Since the original rlswiss contains meta-rl methods and redundant codes, in this repo, we clean and optimize the code architecture, modify and re-implement algorithms for the purpose of easier running imitation learning experiments (rlkit focus on general RL algorithms). We further introduce vec envs to sample data in a parallel style to boost the sampling stage refering to tianshou and add tensorboard support.
You can easily build experiment codes under this framework in your research. We will continue to maintain this repo while keeping it clear and clean.
- Soft-Actor-Critic (SAC)
- Soft-Actor-Critic (SAC) (Auto Learning Alpha version)
- Soft Q Learning (SQL)
- TD3
- DDPG
- PPO
- HER (Goal-Condtioned RL, with SAC or TD3)
- Adversarial Inverse Reinforcement Learning
- AIRL / GAIL / FAIRL / Discriminator-Actor-Critic (DAC) (Different reward signals for AIRL / GAIL / FAIRL, and absorbing state for DAC)
- Behaviour Cloning (bc)
- DAgger
Before running, assign important log and output paths in \rlkit\launchers\common.py
(There is an example file show necessary variables).
Their are simple multiple processing shcheduling (we use multiple processing to clarify it with multi-processing since it only starts many independent sub-process without communication) for simple hyperparameter grid search.
The main entry is run_experiments.py, with the assigned experiment yaml file in \exp_specs
:
python run_experiment.py -g 0 -e your_yaml_path
or CUDA_VISIBLE_DEVICES=0 python run_experiment.py -e your_yaml_path
.
When you run the run_experiments.py, it reads the yaml file, and generate small yaml files with only one hyperparameter setting for each. In a yaml file, a script file path is assigned (see \run_scripts\
), which is specified to run the script with every the small yaml file. See \exp_specs\sac\bc.yaml
for necessary explaination of each parameter.
NOTE: all experiments, including the evaluation tasks (see \run_scripts\evaluate_policy.py
and \exp_specs\evaluate_policy
) and the render tasks, can be run under this framework by specifying the yaml file (in a multiple processes style).
RL algorithms do not need demonstrations. Therefore, all you need is to write an experiment yaml file (see an example in \exp_specs\sac\sac_hopper.yaml
) and run with the above suggestions.
For on-policy algorithms (e.g., PPO), we clean the buffer after every training step.
run sac-ae for finger_spin:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\sac\sac_ae_dmc_finger_spin.yaml
run sac for hopper:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\sac\sac_hopper.yaml
run ppo for hopper:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\ppo\ppo_hopper.yaml
run td3 for humanoid:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\td3\td3_humanoid.yaml
run her for pick with td3:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\her\her_pick_td3.yaml
IL algorithms need to be assigned with demonstrations. A input-format-matching standard mujoco demonstrations can be download in here. If you want to sample your own data, train an expert agent using RL algorithms and sample using \run_scripts\gen_expert_demo.py
or \run_scripts\evaluate_policy.py
, and do not forget to modify your IO format.
If you get the demos ready, write the path for each expert name in demos_listing.yaml
(there are already some examples). Then you should specify the expert name and the traj number in the corresponding yaml file (see \exp_specs\bc.yaml
for example). After all the stuff, you can run it as a regular experiment following the above suggestions.
gen expert data for hopper:
python run_experiment -e \exp_specs\gen_expert\hopper.yaml
run bc for hopper:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\bc.yaml
run gail for walker:
CUDA_VISIBLE_DEVICES=0 python run_experiment -e \exp_specs\gail\gail_walker.yaml
Envs | Mean | Std |
---|---|---|
Pendulum | 139.7313 | 79.8126 |
InvertedPendulum-v2 | 1000.0000 | 0.0000 |
InvertedDoublePendulum-v2 | 9358.8740 | 0.1043 |
Ant-v2 | 5404.5532 | 1520.4961 |
Hopper-v2 | 3402.9494 | 446.4877 |
Humanoid-v2 | 6043.9907 | 726.1788 |
HalfCheetah-v2 | 13711.6445 | 111.4709 |
Walker2d-v2 | 5639.3267 | 29.9715 |
Envs | Mean | Std |
---|---|---|
Finger_Spin (600K) | 983.4286 | 5.8274 |
Envs | Mean | Std |
---|---|---|
InvertedPendulum-v2 | 25.2800 | 5.5318 |
InvertedDoublePendulum-v2 | 78.2829 | 10.7335 |
Ant-v2 | 713.5986 | 203.9204 |
Hopper-v2 | 13.0901 | 0.1022 |
Humanoid-v2 | 64.7384 | 2.3037 |
HalfCheetah-v2 | 74.4849 | 12.3917 |
Walker2d-v2 | 7.0708 | 0.1292 |
Swimmer-v2 | 15.5430 | 6.6655 |