From b3aff441ff357ea10b4b6ee8392744d0c8d48057 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Quentin=20Gallou=C3=A9dec?= <45557362+qgallouedec@users.noreply.github.com> Date: Thu, 12 Dec 2024 16:40:28 +0100 Subject: [PATCH] =?UTF-8?q?=F0=9F=8E=9E=EF=B8=8F=20Add=20"Fine-tuning=20op?= =?UTF-8?q?en=20AI=20models=20using=20Hugging=20Face=20TRL"=20YouTube=20vi?= =?UTF-8?q?deo=20to=20community=20tutorials=20(#2467)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docs/source/community_tutorials.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/docs/source/community_tutorials.md b/docs/source/community_tutorials.md index eb0c5067a3..6aaf03f1b3 100644 --- a/docs/source/community_tutorials.md +++ b/docs/source/community_tutorials.md @@ -11,6 +11,8 @@ Community tutorials are made by active members of the Hugging Face community tha | Preference Optimization | [`DPOTrainer`] | Align Mistral-7b using Direct Preference Optimization for human preference alignment | [Maxime Labonne](https://huggingface.co/mlabonne) | [Link](https://mlabonne.github.io/blog/posts/Fine_tune_Mistral_7b_with_DPO.html) | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/mlabonne/llm-course/blob/main/Fine_tune_a_Mistral_7b_model_with_DPO.ipynb) | | Preference Optimization | [`ORPOTrainer`] | Fine-tuning Llama 3 with ORPO combining instruction tuning and preference alignment | [Maxime Labonne](https://huggingface.co/mlabonne) | [Link](https://mlabonne.github.io/blog/posts/2024-04-19_Fine_tune_Llama_3_with_ORPO.html) | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1eHNWg9gnaXErdAa8_mcvjMupbSS6rDvi) | + + # Vision Language Models | Task | Class | Description | Author | Tutorial | Colab |