Status: Archive (code is provided as-is, no updates expected)
This is code for the environments used in the paper Quantifying Generalization in Reinforcement Learning along with an example training script.
Authors: Karl Cobbe, Oleg Klimov, Chris Hesse, Taehoon Kim, John Schulman
You should install the package in development mode so you can easily change the files. You may also want to create a virtualenv before installing since it depends on a specific version of OpenAI baselines.
This environment has been used on Mac OS X and Ubuntu 16.04 with Python 3.6.
# Linux
apt-get install mpich build-essential qt5-default pkg-config
# Mac
brew install qt open-mpi pkg-config
git clone https://github.com/openai/coinrun.git
cd coinrun
pip install tensorflow==1.12.0 # or tensorflow-gpu
pip install -r requirements.txt
pip install -e .
Note that this does not compile the environment, the environment will be compiled when the coinrun
package is imported.
Try the environment out with the keyboard:
python -m coinrun.interactive
If this fails, you may be missing a dependency or may need to fix coinrun/Makefile
for your machine.
Train an agent using PPO:
python -m coinrun.train_agent --run-id myrun --save-interval 1
After each parameter update, this will save a copy of the agent to ./saved_models/
. Results are logged to /tmp/tensorflow
by default.
Run parallel training using MPI:
mpiexec -np 8 python -m coinrun.train_agent --run-id myrun
Train an agent on a fixed set of N levels:
python -m coinrun.train_agent --run-id myrun --num-levels N
Train an agent on the same 500 levels used in the paper:
python -m coinrun.train_agent --run-id myrun --num-levels 500
Train an agent on a different set of 500 levels:
python -m coinrun.train_agent --run-id myrun --num-levels 500 --set-seed 13
Continue training an agent from a checkpoint:
python -m coinrun.train_agent --run-id newrun --restore-id myrun
Evaluate an agent's final training performance across N parallel environments. Evaluate K levels on each environment (NxK total levels). Default N=20 is reasonable. Evaluation levels will be drawn from the same set as those seen during training.
python -m coinrun.enjoy --train-eval --restore-id myrun -num-eval N -rep K
Evaluate an agent's final test performance on PxNxK distinct levels. All evaluation levels are uniformly sampled from the set of all high difficulty levels. Although we don't explicitly enforce that the test set avoid training levels, the probability of collisions is negligible.
mpiexec -np P python -m coinrun.enjoy --test-eval --restore-id myrun -num-eval N -rep K
Run simultaneous training and testing using MPI. Half the workers will train and half will test.
mpiexec -np 8 python -m coinrun.train_agent --run-id myrun --test
View training options:
python -m coinrun.train_agent --help
Watch a trained agent play a level:
python -m coinrun.enjoy --restore-id myrun --hres
Train an agent to play RandomMazes:
python train_agent.py --run-id random_mazes --game-type maze --use-lstm 1
Train an agent to play CoinRun-Platforms, using a larger number of environments to stabilize learning:
python train_agent.py --run-id coinrun_plat --game-type platform --num-envs 96 --use-lstm 1
There's an example random agent script in coinrun/random_agent.py
which you can run like so:
python -m coinrun.random_agent
There's also a Dockerfile
to create a CoinRun docker image:
docker build --tag coinrun .
docker run --rm coinrun python -um coinrun.random_agent
There's also a Colab notebook showing how to setup CoinRun.
See LICENSES for asset license information.