Skip to content

Commit 28058d1

Browse files
committed
update readme
1 parent 282a721 commit 28058d1

File tree

1 file changed

+9
-16
lines changed

1 file changed

+9
-16
lines changed

examples/llama_3.2_1b/README.md

Lines changed: 9 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -12,34 +12,27 @@ To download the necessary files for the model, please follow the links below:
1212
- **Model File**: [model.safetensors](https://huggingface.co/meta-llama/Llama-3.2-1B/tree/main)
1313
- **Tokenizer File**: [tokenizer.model](https://huggingface.co/meta-llama/Llama-3.2-1B/tree/main/original)
1414

15-
Make sure to place these files in the appropriate directory for your project.
15+
The CI tests expect these files to be placed in the directory `/srv/llama3.2-1b`;
16+
when calling `inference.py`, you will supply the paths to those two files on the command line, so you can place them anywhere.
1617

1718
## Installation Instructions
1819

1920
Before running `inference.py`, ensure you have the proper environment. To build the environment from scratch, follow the instructions below:
2021

21-
1. Create and activate a Python virtual environment:
22-
```bash
23-
python3 -m venv ironenv
24-
source ironenv/bin/activate
25-
python3 -m pip install --upgrade pip
26-
```
22+
1. Follow the IRON installation instructions in the repository root fist.
23+
After this, you should have an `ironenv` environment set up and activated.
2724

28-
2. Install prerequisites:
29-
```bash
30-
source ./scripts/install_prereqs.sh --env <name of your virtual env>
31-
source ./script/env_setup.sh --env <name of your virtual env>
32-
python3 -m pip install -r requirements_llama.txt
33-
```
34-
35-
You can skip passing the `--env` option if you named your virtual environment `ironenv`.
25+
2. Install the following additional requirements:
26+
```
27+
python3 -m pip install -r requirements_examples.txt
28+
```
3629

3730
## Running Inference
3831

3932
Inference with Llama-3.2-1B can be run by specifying a number of tokens to generate based on a prompt. This is done with `inference.py`:
4033
```bash
4134
cd golden_model
42-
python inference.py model.safetensors tokenizer.model --num_tokens <NUM_TOKENS> --prompt <PROMPT>
35+
python inference.py /path/to/model.safetensors /path/to/tokenizer.model --num_tokens <NUM_TOKENS> --prompt <PROMPT>
4336
```
4437

4538
`inference.py` has the following command format:

0 commit comments

Comments
 (0)