Skip to content
This repository has been archived by the owner on Jun 27, 2023. It is now read-only.
/ hiveformer-corl Public archive

PyTorch implementation of the Hiveformer research paper

License

Notifications You must be signed in to change notification settings

vlc-robot/hiveformer-corl

Repository files navigation

🐝 Hiveformer 🐝

⚠️ WARNING

The repository is not maintained anymore. You can find a reimplementation of the paper on this repository.

Licence PyTorch arXiv Star this repository GitHub Actions

This is the PyTorch implementation of the Hiveformer research paper:

Instruction-driven history-aware policies for robotic manipulations
Pierre-Louis Guhur, Shizhe Chen, Ricardo Garcia, Makarand Tapaswi, Ivan Laptev, Cordelia Schmid
CoRL 2022 (oral)

🛠️ 1. Getting started

Clone the repository along with its submodules:

git clone --recursive https://github.com/guhur/hiveformer

# or
git clone https://github.com/guhur/hiveformer
git submodule --init --recursive update

You need a recent version of Python (higher or equal than 3.9) and install dependencies:

poetry install

# this should be run on every shell
poetry shell

Other dependencies (RLBench, PyRep) need to be installed manually. Launch corresponding Makefile rules:

make pyrep
make rlbench

Note that you can also use a Docker or Singularity container:

make container

💽 2. Preparing dataset

Hiveformer is training over an offline dataset of succesful demonstrations. You can generate demonstrations from the corresponding SLURM file or with the following command:

data_dir=/path/to/raw/dataset
output_dir=/path/to/packaged/dataset
for seed in 0 1 2 3 4; do
  cd /path/to/rlbench
  # Generate samples
  python dataset_generator.py \
    --save_path=$data_dir/$seed \
    --tasks=$(cat tasks.csv | tr '\n' ' ') \
    --image_size=128,128 \
    --renderer=opengl \
    --episodes_per_task=100 \
    --variations=1 \
    --offset=0 \
    --processes=1

  cd /path/to/hiveformer
  for task in $(cat tasks.csv | tr '\n' ' '); do
  python data_gen.py \
   --data_dir=$data_dir/$seed \
   --output=$output_dir \
   --max_variations=1 \
   --num_episodes=100 \
   --tasks=$task \
  done
done

Next, you need to preprocess instructions:

python preprocess_instructions.py \
	--tasks $(cat tasks.csv | tr '\n' ' ')
	--output instructions.pkl \
	--variations {0..199} \
	--annotations annotations.json

🏋️ 3. Train your agents

3.1. Single-task learning

for seed in 0 1 2 3 4 5; do
for task in $(cat tasks.csv | tr '\n' ' '); do
python train.py \
	--tasks $task \
	--dataset $output_dir/$seed \
	--num_workers 10  \
 	--instructions instructions.pkl \
	--variations 0
done
done

3.2. Multi-task learning

for seed in 0 1 2 3 4; do
python train.py \
	--tasks $(cat tasks.csv | tr '\n' ' ') \
	--dataset $output_dir/$seed \
	--num_workers 10  \
 	--instructions instructions.pkl \
	--variations 0
done

3.3. Multi-variation learning

for seed in 0 1 2 3 4; do
for task in push_buttons tower3; do
python train.py \
	--arch mct \
	--tasks $task \
	--dataset $output_dir/$seed \
	--num_workers 10  \
 	--instructions instructions.pkl \
	--variations {0..99}
done
done

4. ⏱️ Evaluation

python eval.py \
	--checkpoint /path/to/checkpoint/ 
	--variations 0 \
	--instructions instructions.pkl \
	--num_episodes 100

🙏 Credits

Parts of the code were copied from Auto-Lambda, Cosy Pose, LXMERT and RLBench.

About

PyTorch implementation of the Hiveformer research paper

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published