Skip to content

Commit 5c6a67a

Browse files
committed
Keep faqs consistent with main
Signed-off-by: wangli <wangli858794774@gmail.com>
1 parent 555fc9b commit 5c6a67a

File tree

1 file changed

+18
-8
lines changed

1 file changed

+18
-8
lines changed

docs/source/faqs.md

Lines changed: 18 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@
55
- [[v0.7.1rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/19)
66
- [[v0.7.3rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/267)
77
- [[v0.7.3rc2] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/418)
8+
- [[v0.8.4rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/546)
89

910
## General FAQs
1011

@@ -22,21 +23,21 @@ Below series are NOT supported yet:
2223

2324
From a technical view, vllm-ascend support would be possible if the torch-npu is supported. Otherwise, we have to implement it by using custom ops. We are also welcome to join us to improve together.
2425

25-
2626
### 2. How to get our docker containers?
2727

2828
You can get our containers at `Quay.io`, e.g., [<u>vllm-ascend</u>](https://quay.io/repository/ascend/vllm-ascend?tab=tags) and [<u>cann</u>](https://quay.io/repository/ascend/cann?tab=tags).
2929

3030
If you are in China, you can use `daocloud` to accelerate your downloading:
3131

3232
```bash
33-
TAG=vllm-ascend released tag you want to pull(eg: v0.7.3rc2)
33+
# Replace with tag you want to pull
34+
TAG=v0.7.3rc2
3435
docker pull m.daocloud.io/quay.io/ascend/vllm-ascend:$TAG
3536
```
3637

3738
### 3. What models does vllm-ascend supports?
3839

39-
Find more details [<u>here</u>](https://vllm-ascend.readthedocs.io/en/v0.7.3-dev/user_guide/supported_models.html).
40+
Find more details [<u>here</u>](https://vllm-ascend.readthedocs.io/en/latest/user_guide/supported_models.html).
4041

4142
### 4. How to get in touch with our community?
4243

@@ -77,24 +78,33 @@ vllm-ascend is a plugin for vllm. Basically, the version of vllm-ascend is the s
7778

7879
### 9. Does vllm-ascend support Prefill Disaggregation feature?
7980

80-
Currently, only 1P1D is supported by vllm. For vllm-ascend, it'll be done by [this PR](https://github.com/vllm-project/vllm-ascend/pull/539). For NPND, vllm is not stable and fully supported yet. We will make it stable and supported by vllm-ascend in the future.
81+
Currently, only 1P1D is supported by vllm. For vllm-ascend, it'll be done by [this PR](https://github.com/vllm-project/vllm-ascend/pull/432). For NPND, vllm is not stable and fully supported yet. We will make it stable and supported by vllm-ascend in the future.
8182

8283
### 10. Does vllm-ascend support quantization method?
8384

84-
Currently, w8a8 quantization is supporeted with the integration of vllm-ascend and mindie-turbo, please use `pip install vllm-ascend[mindie-turbo]`
85+
Currently, w8a8 quantization is already supported by vllm-ascend originally on v0.8.4rc2 or heigher, If you're using vllm 0.7.3 version, w8a8 quantization is supporeted with the integration of vllm-ascend and mindie-turbo, please use `pip install vllm-ascend[mindie-turbo]`.
86+
87+
### 11. How to run w8a8 DeepSeek model?
88+
89+
Currently, w8a8 DeepSeek is working in process: [support AscendW8A8 quantization](https://github.com/vllm-project/vllm-ascend/pull/511)
8590

86-
### 11. There is not output in log when loading models using vllm-ascend, How to solve it?
91+
Please run DeepSeek with BF16 now, follwing the [Multi-Node DeepSeek inferencing tutorail](https://vllm-ascend.readthedocs.io/en/main/tutorials/multi_node.html)
92+
93+
### 12. There is not output in log when loading models using vllm-ascend, How to solve it?
8794

8895
If you're using vllm 0.7.3 version, this is a known progress bar display issue in VLLM, which has been resolved in [this PR](https://github.com/vllm-project/vllm/pull/12428), please cherry-pick it locally by yourself. Otherwise, please fill up an issue.
8996

90-
### 12. How vllm-ascend is tested
97+
### 13. How vllm-ascend is tested
9198

9299
vllm-ascend is tested by functional test, performance test and accuracy test.
93100

94-
- **Functional test**: we added CI, includes portion of vllm's native unit tests and vllm-ascend's own unit tests,on vllm-ascend's test, we test basic functionality and [supported features](https://vllm-ascend.readthedocs.io/en/v0.7.3-dev/user_guide/suppoted_features.html)
101+
- **Functional test**: we added CI, includes portion of vllm's native unit tests and vllm-ascend's own unit tests,on vllm-ascend's test, we test basic functionality、popular models availability and [supported features](https://vllm-ascend.readthedocs.io/en/latest/user_guide/suppoted_features.html) via e2e test
95102

96103
- **Performance test**: we provide [benchmark](https://github.com/vllm-project/vllm-ascend/tree/main/benchmarks) tools for end-to-end performance benchmark which can easily to re-route locally, we'll publish a perf website like [vllm](https://simon-mo-workspace.observablehq.cloud/vllm-dashboard-v0/perf) does to show the performance test results for each pull request
97104

98105
- **Accuracy test**: we're working on adding accuracy test to CI as well.
99106

100107
Finnall, for each release, we'll publish the performance test and accuracy test report in the future.
108+
109+
### 14. How to fix the error "InvalidVersion" when using vllm-ascend?
110+
It's usually because you have installed an dev/editable version of vLLM package. In this case, we provide the env variable `VLLM_VERSION` to let users specify the version of vLLM package to use. Please set the env variable `VLLM_VERSION` to the version of vLLM package you have installed. The format of `VLLM_VERSION` should be `X.Y.Z`.

0 commit comments

Comments
 (0)