|
9 | 9 |
|
10 | 10 | [**Introduction**](#introduction) | [**Installation**](#installation) | [**Get Started**](#get-started) | [**Documentation**](https://pytorch.org/torchtune/main/index.html) | [**Community**](#community) | [**License**](#license) | [**Citing torchtune**](#citing-torchtune) |
11 | 11 |
|
12 | | -> [!IMPORTANT] |
13 | | -> Update September 25, 2024: torchtune has support for **Llama 3.2 11B Vision**, **Llama 3.2 3B**, and **Llama 3.2 1B** models! Try them out by following our installation instructions [here](#Installation), then run any of the text configs [here](recipes/configs/llama3_2) or vision configs [here](recipes/configs/llama3_2_vision). |
| 12 | +### 📣 Recent updates 📣 |
| 13 | +* *November 2024*: torchtune has released [v0.4.0](https://github.com/pytorch/torchtune/releases/tag/v0.4.0) which includes stable support for exciting features like activation offloading and multimodal QLoRA |
| 14 | +* *November 2024*: torchtune has added [Gemma2](recipes/configs/gemma2) to its models! |
| 15 | +* *October 2024*: torchtune added support for Qwen2.5 models - find the recipes [here](recipes/configs/qwen2_5/) |
| 16 | +* *September 2024*: torchtune has support for **Llama 3.2 11B Vision**, **Llama 3.2 3B**, and **Llama 3.2 1B** models! Try them out by following our installation instructions [here](#Installation), then run any of the text configs [here](recipes/configs/llama3_2) or vision configs [here](recipes/configs/llama3_2_vision). |
14 | 17 |
|
15 | 18 |
|
16 | 19 | |
@@ -44,8 +47,10 @@ torchtune currently supports the following models. |
44 | 47 | | [Code-Llama2](https://ai.meta.com/blog/code-llama-large-language-model-coding/) | 7B, 13B, 70B [[models](torchtune/models/code_llama2/_model_builders.py), [configs](recipes/configs/code_llama2/)] | |
45 | 48 | | [Mistral](https://huggingface.co/mistralai) | 7B [[models](torchtune/models/mistral/_model_builders.py), [configs](recipes/configs/mistral/)] | |
46 | 49 | | [Gemma](https://huggingface.co/collections/google/gemma-release-65d5efbccdbb8c4202ec078b) | 2B, 7B [[models](torchtune/models/gemma/_model_builders.py), [configs](recipes/configs/gemma/)] | |
| 50 | +| [Gemma2](https://huggingface.co/docs/transformers/main/en/model_doc/gemma2) | 2B, 9B, 27B [[models](torchtune/models/gemma2/_model_builders.py), [configs](recipes/configs/gemma2/)] | |
47 | 51 | | [Microsoft Phi3](https://huggingface.co/collections/microsoft/phi-3-6626e15e9585a200d2d761e3) | Mini [[models](torchtune/models/phi3/), [configs](recipes/configs/phi3/)] |
48 | 52 | | [Qwen2](https://qwenlm.github.io/blog/qwen2/) | 0.5B, 1.5B, 7B [[models](torchtune/models/qwen2/), [configs](recipes/configs/qwen2/)] |
| 53 | +| [Qwen2.5](https://qwenlm.github.io/blog/qwen2.5/) | 0.5B, 1.5B, 3B, 7B, 14B, 32B, 72B [[models](torchtune/models/qwen2_5/), [configs](recipes/configs/qwen2_5/)] |
49 | 54 |
|
50 | 55 | We're always adding new models, but feel free to [file an issue](https://github.com/pytorch/torchtune/issues/new) if there's a new one you would like to see in torchtune. |
51 | 56 |
|
@@ -162,6 +167,7 @@ To download Llama3.1, you can run: |
162 | 167 | ```bash |
163 | 168 | tune download meta-llama/Meta-Llama-3.1-8B-Instruct \ |
164 | 169 | --output-dir /tmp/Meta-Llama-3.1-8B-Instruct \ |
| 170 | +--ignore-patterns "original/consolidated.00.pth" \ |
165 | 171 | --hf-token <HF_TOKEN> \ |
166 | 172 | ``` |
167 | 173 |
|
@@ -258,6 +264,7 @@ We really value our community and the contributions made by our wonderful users. |
258 | 264 | - [@fyabc](https://github.com/fyabc) for adding Qwen2 models, tokenizer, and recipe integration to torchtune |
259 | 265 | - [@solitude-alive](https://github.com/solitude-alive) for adding the [Gemma 2B model](torchtune/models/gemma/) to torchtune, including recipe changes, numeric validations of the models and recipe correctness |
260 | 266 | - [@yechenzhi](https://github.com/yechenzhi) for adding [Direct Preference Optimization (DPO)](recipes/lora_dpo_single_device.py) to torchtune, including the recipe and config along with correctness checks |
| 267 | +- [@Optimox](https://github.com/Optimox) for adding all the [Gemma2 variants](torchtune/models/gemma2) to torchtune! |
261 | 268 |
|
262 | 269 |
|
263 | 270 | |
|
0 commit comments