Skip to content

Latest commit

 

History

History
9 lines (5 loc) · 648 Bytes

README.md

File metadata and controls

9 lines (5 loc) · 648 Bytes

Finetuning quantized models

Finetune by emulating 8bit GEMM during training. Student models are more difficult to quantize, so you should finetune them to reduce the BLEU hit.

Example is shown in run.me.finetune.example.sh. Finetuning is really fast. The model's quality is going to start going down after a few thousand mini-batches. Make sure you have frequent validations so that you don't miss the sweet spot!