This repository contains a series of works for representional learning in RL.
The main branch contains the code for Briee, the ICML 2022 paper Efficient Reinforcement Learning in Block MDPs: A Model-free Representation Learning approach.
The reptransfer branch contains the code for RepTransfer, Provable Benefits of Representational Transfer in Reinforcement Learning
Creating a virtual environment is recommended (or using conda alternatively):
pip install virtualenv
virtualenv /path/to/venv --python=python3
#To activate a virtualenv:
. /path/to/venv/bin/activate
To install the dependencies (the results from the paper are obtain from gym==0.14.0):
pip install -r requirements.txt
To install pytorch, please follow PyTorch. Note that the current implementation does not require pytorch gpu.
We use wandb to perform result collection, please setup wandb before running the code or add os.environ['WANDB_MODE'] = 'offline'
in main.py
.
To reproduce our result in comblock, please run:
bash run.sh [horizon] [num_threads] [save_path]
To reproduce our result in comblock with simplex feature, please run:
bash run_simplex.sh [num_threads] [save_path]
To reproduce our result in comblock with dense reward, please run:
bash run_dense.sh [num_threads] [save_path]
To see all the hyperparameters, please refer to utils.py
.
Please refer to this repo for reproducing PPO+RND results.