@@ -12,34 +12,27 @@ To download the necessary files for the model, please follow the links below:
1212- ** Model File** : [ model.safetensors] ( https://huggingface.co/meta-llama/Llama-3.2-1B/tree/main )
1313- ** Tokenizer File** : [ tokenizer.model] ( https://huggingface.co/meta-llama/Llama-3.2-1B/tree/main/original )
1414
15- Make sure to place these files in the appropriate directory for your project.
15+ The CI tests expect these files to be placed in the directory ` /srv/llama3.2-1b ` ;
16+ when calling ` inference.py ` , you will supply the paths to those two files on the command line, so you can place them anywhere.
1617
1718## Installation Instructions
1819
1920Before running ` inference.py ` , ensure you have the proper environment. To build the environment from scratch, follow the instructions below:
2021
21- 1 . Create and activate a Python virtual environment:
22- ``` bash
23- python3 -m venv ironenv
24- source ironenv/bin/activate
25- python3 -m pip install --upgrade pip
26- ```
22+ 1 . Follow the IRON installation instructions in the repository root fist.
23+ After this, you should have an ` ironenv ` environment set up and activated.
2724
28- 2. Install prerequisites:
29- ` ` ` bash
30- source ./scripts/install_prereqs.sh --env < name of your virtual env>
31- source ./script/env_setup.sh --env < name of your virtual env>
32- python3 -m pip install -r requirements_llama.txt
33- ` ` `
34-
35- You can skip passing the ` --env` option if you named your virtual environment ` ironenv` .
25+ 2 . Install the following additional requirements:
26+ ```
27+ python3 -m pip install -r requirements_examples.txt
28+ ```
3629
3730## Running Inference
3831
3932Inference with Llama-3.2-1B can be run by specifying a number of tokens to generate based on a prompt. This is done with ` inference.py ` :
4033``` bash
4134cd golden_model
42- python inference.py model.safetensors tokenizer.model --num_tokens < NUM_TOKENS> --prompt < PROMPT>
35+ python inference.py /path/to/ model.safetensors /path/to/ tokenizer.model --num_tokens < NUM_TOKENS> --prompt < PROMPT>
4336```
4437
4538` inference.py ` has the following command format:
0 commit comments