You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -39,7 +39,7 @@ By using vLLM Ascend plugin, popular open-source models, including Transformer-l
39
39
* PyTorch >= 2.4.0, torch-npu >= 2.4.0
40
40
* vLLM (the same version as vllm-ascend)
41
41
42
-
Find more about how to setup your environment step by step in [here](docs/installation.md).
42
+
Find more about how to setup your environment step by step in [here](docs/source/installation.md).
43
43
44
44
## Getting Started
45
45
@@ -68,7 +68,7 @@ Run the following command to start the vLLM server with the [Qwen/Qwen2.5-0.5B-I
68
68
vllm serve Qwen/Qwen2.5-0.5B-Instruct
69
69
curl http://localhost:8000/v1/models
70
70
```
71
-
**Please refer to [official docs](./docs/index.md) for more details.**
71
+
**Please refer to [official docs](https://vllm-ascend.readthedocs.io/en/latest/) for more details.**
72
72
73
73
## Contributing
74
74
See [CONTRIBUTING](docs/source/developer_guide/contributing.md) for more details, which is a step-by-step guide to help you set up development environment, build and test.
| Python | >= 3.9 |[3.10](https://www.python.org/downloads/)| Required for vllm |
7
-
| CANN | >= 8.0.RC2 |[8.0.RC3](https://www.hiascend.com/developer/download/community/result?module=cann&cann=8.0.0.beta1)| Required for vllm-ascend and torch-npu |
8
-
| torch-npu | >= 2.4.0 |[2.5.1rc1](https://gitee.com/ascend/pytorch/releases/tag/v6.0.0.alpha001-pytorch2.5.1)| Required for vllm-ascend |
9
-
| torch | >= 2.4.0 |[2.5.1](https://github.com/pytorch/pytorch/releases/tag/v2.5.1)| Required for torch-npu and vllm required |
3
+
This document describes how to install vllm-ascend manually.
10
4
11
-
## Prepare Ascend NPU environment
5
+
## Requirements
12
6
13
-
Below is a quick note to install recommended version software:
7
+
- OS: Linux
8
+
- Python: 3.10 or higher
9
+
- A hardware with Ascend NPU. It's usually the Atlas 800 A2 series.
10
+
- Software:
14
11
15
-
### Containerized installation
12
+
| Software | Supported version | Note |
13
+
| ------------ | ----------------- | ---- |
14
+
| CANN | >= 8.0.0.beta1 | Required for vllm-ascend and torch-npu |
15
+
| torch-npu | >= 2.5.1rc1 | Required for vllm-ascend |
16
+
| torch | >= 2.5.1 | Required for torch-npu and vllm |
16
17
17
-
You can use the [container image](https://hub.docker.com/r/ascendai/cann) directly with one line command:
18
+
## Configure a new environment
19
+
20
+
Before installing the package, you need to make sure that the firmware and driver for NPU is installed correctly. i.e. `npu-smi` command is available.
21
+
22
+
> Tips: following the instructions provided in the [Ascend Installation Guide](https://ascend.github.io/docs/sources/ascend/quick_install.html) can help you to set up the environment easily.
23
+
24
+
Once it's done, you can read either **Set up using Python** or **Set up using Docker** section to install and use vllm-ascend.
25
+
26
+
If you want to install vllm-ascend on a bare environment by hand, you need install CANN first, otherwise, you can skip this step if using container environment.
Or follow the instructions provided in the [Ascend Installation Guide](https://ascend.github.io/docs/sources/ascend/quick_install.html) to set up the environment.
0 commit comments