Skip to content

Commit 75e2005

Browse files
PotabkAngazenn
authored andcommitted
[Doc] Update faqs (vllm-project#699)
### What this PR does / why we need it? Update faqs to make it more clear Signed-off-by: wangli <wangli858794774@gmail.com>
1 parent 279fd17 commit 75e2005

File tree

1 file changed

+5
-3
lines changed

1 file changed

+5
-3
lines changed

docs/source/faqs.md

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,9 @@ You can get our containers at `Quay.io`, e.g., [<u>vllm-ascend</u>](https://quay
3030
If you are in China, you can use `daocloud` to accelerate your downloading:
3131

3232
```bash
33-
docker pull m.daocloud.io/quay.io/ascend/vllm-ascend:v0.7.3rc2
33+
# Replace with tag you want to pull
34+
TAG=v0.7.3rc2
35+
docker pull m.daocloud.io/quay.io/ascend/vllm-ascend:$TAG
3436
```
3537

3638
### 3. What models does vllm-ascend supports?
@@ -80,7 +82,7 @@ Currently, only 1P1D is supported by vllm. For vllm-ascend, it'll be done by [th
8082

8183
### 10. Does vllm-ascend support quantization method?
8284

83-
Currently, there is no quantization method supported in vllm-ascend originally. And the quantization supported is working in progress, w8a8 will firstly be supported.
85+
Currently, w8a8 quantization is already supported by vllm-ascend originally on v0.8.4rc2 or heigher, If you're using vllm 0.7.3 version, w8a8 quantization is supporeted with the integration of vllm-ascend and mindie-turbo, please use `pip install vllm-ascend[mindie-turbo]`.
8486

8587
### 11. How to run w8a8 DeepSeek model?
8688

@@ -96,7 +98,7 @@ If you're using vllm 0.7.3 version, this is a known progress bar display issue i
9698

9799
vllm-ascend is tested by functional test, performance test and accuracy test.
98100

99-
- **Functional test**: we added CI, includes portion of vllm's native unit tests and vllm-ascend's own unit tests,on vllm-ascend's test, we test basic functional usability for popular models, include `Qwen2.5-7B-Instruct``Qwen2.5-VL-7B-Instruct``Qwen2.5-VL-32B-Instruct``QwQ-32B`.
101+
- **Functional test**: we added CI, includes portion of vllm's native unit tests and vllm-ascend's own unit tests,on vllm-ascend's test, we test basic functionality、popular models availability and [supported features](https://vllm-ascend.readthedocs.io/en/latest/user_guide/suppoted_features.html) via e2e test
100102

101103
- **Performance test**: we provide [benchmark](https://github.com/vllm-project/vllm-ascend/tree/main/benchmarks) tools for end-to-end performance benchmark which can easily to re-route locally, we'll publish a perf website like [vllm](https://simon-mo-workspace.observablehq.cloud/vllm-dashboard-v0/perf) does to show the performance test results for each pull request
102104

0 commit comments

Comments
 (0)