Skip to content

Commit

Permalink
Merge pull request #7 from alquist4121/fix-example-nllb200_3B_8bit
Browse files Browse the repository at this point in the history
FIX: update explanation text for nllb200_3B_8bit.sh
  • Loading branch information
ikergarcia1996 authored Jul 31, 2023
2 parents bb568bf + b8ba25f commit f88323f
Showing 1 changed file with 5 additions and 2 deletions.
7 changes: 5 additions & 2 deletions examples/nllb200_3B_8bit.sh
Original file line number Diff line number Diff line change
@@ -1,5 +1,8 @@
# Run NLLB200-3B model on sample text. We use FP16 precision, which requires a GPU with a lot of VRAM
# For running this model in GPUs with less VRAM, use 8-bit quantization, see examples/nllb200_3B_8bit.sh
# Run NLLB200-3B on sample text. This model requires a GPU with a lot of VRAM, so we use
# 8-bit quantization to reduce the required VRAM so we can fit in customer grade GPUs. If you have a GPU
# with a lot of RAM, running the model in FP16 should be faster and produce sighly better results,
# see examples/nllb200-3B_fp16.sh


python3 translate.py \
--sentences_path sample_text/en.txt \
Expand Down

0 comments on commit f88323f

Please sign in to comment.