Repo for the paper Symbolic Chain of Thought Distillation.
pip install transformers==4.21.1 tensorboard yacs tensorboardX accelerate==0.14.0 pandas wandb openai sentencepiece datasets==1.18.3 torchtext deepspeed matplotlib seaborn markupsafe==2.0.1 sentence_transformers fastcluster
Please run accelerate config
to set up the accelerate environment. We used DeepSpeed (default config and ZeRO optimization level 2) and fp16 to train the models on a A6000.
Please download the data and put them under "DATA" folder. Data is available at here.
DATA
csqa_30x.json
quarel_30x.json
openbook_30x.json
export GPU_ID=0
bash _scripts/train_single_task.bash \
$GPU_ID 1720 \
configs/train.yaml \
CSQA_30x configs_data/complete/commonsenseqa.py DATA/csqa_30x.json \
RUN_UID csqa
export GPU_ID=0
bash _scripts/train_single_task.bash \
$GPU_ID 1721 \
configs/train.yaml \
QuaRel_30x configs_data/complete/quarel.py DATA/quarel_30x.json \
RUN_UID quarel
export GPU_ID=0
bash _scripts/train_single_task.bash \
$GPU_ID 1722 \
configs/train.yaml \
OB_30x configs_data/complete/openbook.py DATA/openbook_30x.json \
RUN_UID openbook
export DIR_EVAL=OUTPUTS/{RUN_UID}
export CONFIG=configs_data/complete/{task}.py
export MODEL_EPOCH=best
export GPU_ID=0
bash _scripts/eval_single_task.bash $GPU_ID \
DATA.CONFIG $CONFIG \
TRAIN.SAMPLE_ROUNDS 1 \
POLICY_MODEL.TEMPERATURE 0.0 \
POLICY_MODEL.DO_SAMPLE False
Some pre-trained checkpoints are available at here.