Skip to content

Latest commit

 

History

History
46 lines (29 loc) · 1.63 KB

quant.md

File metadata and controls

46 lines (29 loc) · 1.63 KB

Quantization

Most of the code has an apache license from meta llama.

We are performing post training quantization. After training has been completed, we quantize the layers in order, instead of quantizing everything at once. Will elaborate more about quantization later on. tokenizer.py generates tokenizer.model. Add it to your model.

llama quantization

I haven't included the llama quantization. You can follow the instructions below if you wish to try:

Downloading pretrained weights Except for when you are training from scratch, you will need the pretrained weights from Meta.

Original Meta weights

Download the model weights following the instructions on the official LLaMA repository.

Once downloaded, you should have a folder like this:

checkpoints/llama
├── 7B
│   ├── ...
│   └── consolidated.00.pth
├── 13B
│   ...
└── tokenizer.model

Convert the weights to the LLaMA format:

Convert the checkpoints. ou are all set. Now you can continue with inference or finetuning.

Try creating a generate() as we have done in babygpt_trainer.py to test the imported weights.

Note: All scripts are prone to argument customization

OpenLLaMA

OpenLM Research has released Apache 2.0 licensed weights obtained by training LLaMA on the 1.2 trillion token open-source RedPajama dataset.

Weights were released in preview on intermediate number of tokens (400B at the time of writing). In order to get them do:

Make sure you have git-lfs installed (https://git-lfs.com): git lfs install

git clone https://huggingface.co/openlm-research/open_llama_7b_400bt_preview checkpoints/open-llama/7B