Skip to content

Commit df5c879

Browse files
authored
[doc] update wrong hf model links (#17184)
Signed-off-by: reidliu41 <reid201711@gmail.com> Co-authored-by: reidliu41 <reid201711@gmail.com>
1 parent 423e9f1 commit df5c879

File tree

5 files changed

+6
-7
lines changed

5 files changed

+6
-7
lines changed

docs/source/features/quantization/auto_awq.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ To create a new 4-bit quantized model, you can leverage [AutoAWQ](https://github
66
Quantization reduces the model's precision from BF16/FP16 to INT4 which effectively reduces the total model memory footprint.
77
The main benefits are lower latency and memory usage.
88

9-
You can quantize your own models by installing AutoAWQ or picking one of the [6500+ models on Huggingface](https://huggingface.co/models?sort=trending&search=awq).
9+
You can quantize your own models by installing AutoAWQ or picking one of the [6500+ models on Huggingface](https://huggingface.co/models?search=awq).
1010

1111
```console
1212
pip install autoawq

docs/source/features/quantization/bitblas.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,8 +20,8 @@ vLLM reads the model's config file and supports pre-quantized checkpoints.
2020

2121
You can find pre-quantized models on:
2222

23-
- [Hugging Face (BitBLAS)](https://huggingface.co/models?other=bitblas)
24-
- [Hugging Face (GPTQ)](https://huggingface.co/models?other=gptq)
23+
- [Hugging Face (BitBLAS)](https://huggingface.co/models?search=bitblas)
24+
- [Hugging Face (GPTQ)](https://huggingface.co/models?search=gptq)
2525

2626
Usually, these repositories have a `quantize_config.json` file that includes a `quantization_config` section.
2727

docs/source/features/quantization/bnb.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ pip install bitsandbytes>=0.45.3
1414

1515
vLLM reads the model's config file and supports both in-flight quantization and pre-quantized checkpoint.
1616

17-
You can find bitsandbytes quantized models on <https://huggingface.co/models?other=bitsandbytes>.
17+
You can find bitsandbytes quantized models on <https://huggingface.co/models?search=bitsandbytes>.
1818
And usually, these repositories have a config.json file that includes a quantization_config section.
1919

2020
## Read quantized checkpoint

docs/source/features/quantization/gptqmodel.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ for more details on this and other advanced features.
1818

1919
## Installation
2020

21-
You can quantize your own models by installing [GPTQModel](https://github.com/ModelCloud/GPTQModel) or picking one of the [5000+ models on Huggingface](https://huggingface.co/models?sort=trending&search=gptq).
21+
You can quantize your own models by installing [GPTQModel](https://github.com/ModelCloud/GPTQModel) or picking one of the [5000+ models on Huggingface](https://huggingface.co/models?search=gptq).
2222

2323
```console
2424
pip install -U gptqmodel --no-build-isolation -v

docs/source/features/quantization/torchao.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,5 +30,4 @@ tokenizer.push_to_hub(hub_repo)
3030
quantized_model.push_to_hub(hub_repo, safe_serialization=False)
3131
```
3232

33-
Alternatively, you can use the TorchAO Quantization space for quantizing models with a simple UI.
34-
See: https://huggingface.co/spaces/medmekk/TorchAO_Quantization
33+
Alternatively, you can use the [TorchAO Quantization space](https://huggingface.co/spaces/medmekk/TorchAO_Quantization) for quantizing models with a simple UI.

0 commit comments

Comments
 (0)