diff --git a/doc/Install_CN.md b/doc/Install_CN.md index acff887f2..6584016c3 100644 --- a/doc/Install_CN.md +++ b/doc/Install_CN.md @@ -4,10 +4,16 @@ - [1.使用开发镜像](#1) - [Serving 开发镜像](#1.1) + - [CPU 镜像](#1.1.1) + - [GPU 镜像](#1.1.2) + - [ARM & XPU 镜像](#1.1.3) - [Paddle 开发镜像](#1.2) + - [CPU 镜像](#1.2.1) + - [GPU 镜像](#1.2.2) - [2.安装 Wheel 包](#2) - [在线安装](#2.1) - [离线安装](#2.2) + - [ARM & XPU 包安装](#2.3) - [3.环境检查](#3) @@ -33,33 +39,52 @@ | CUDA10.2 + cuDNN 7 | 0.9.0-cuda10.2-cudnn7-devel | Ubuntu 16 | 2.3.0-gpu-cuda10.2-cudnn7 | Ubuntu 18 | CUDA10.2 + cuDNN 8 | 0.9.0-cuda10.2-cudnn8-devel | Ubuntu 16 | 无 | Ubuntu 18 | | CUDA11.2 + cuDNN 8 | 0.9.0-cuda11.2-cudnn8-devel | Ubuntu 16 | 2.3.0-gpu-cuda11.2-cudnn8 | Ubuntu 18 | +| ARM + XPU | xpu-arm | CentOS 8.3 | 无 | 无 | 对于**Windows 10 用户**,请参考文档[Windows平台使用Paddle Serving指导](Windows_Tutorial_CN.md)。 - ### 1.1 Serving开发镜像(CPU/GPU 2选1) + + + **CPU:** ``` # 启动 CPU Docker docker pull registry.baidubce.com/paddlepaddle/serving:0.9.0-devel -docker run -p 9292:9292 --name test -dit registry.baidubce.com/paddlepaddle/serving:0.9.0-devel bash -docker exec -it test bash +docker run -p 9292:9292 --name test_cpu -dit registry.baidubce.com/paddlepaddle/serving:0.9.0-devel bash +docker exec -it test_cpu bash git clone https://github.com/PaddlePaddle/Serving ``` + + + **GPU:** ``` # 启动 GPU Docker docker pull registry.baidubce.com/paddlepaddle/serving:0.9.0-cuda11.2-cudnn8-devel -nvidia-docker run -p 9292:9292 --name test -dit registry.baidubce.com/paddlepaddle/serving:0.9.0-cuda11.2-cudnn8-devel bash -nvidia-docker exec -it test bash +nvidia-docker run -p 9292:9292 --name test_gpu -dit registry.baidubce.com/paddlepaddle/serving:0.9.0-cuda11.2-cudnn8-devel bash +nvidia-docker exec -it test_gpu bash +git clone https://github.com/PaddlePaddle/Serving +``` + + + +**ARM & XPU: ** +``` +docker pull registry.baidubce.com/paddlepaddle/serving:xpu-arm +docker run -p 9292:9292 --name test_arm_xpu -dit registry.baidubce.com/paddlepaddle/serving:xpu-arm bash +docker exec -it test_arm_xpu bash git clone https://github.com/PaddlePaddle/Serving ``` ### 1.2 Paddle开发镜像(CPU/GPU 2选1) + + + **CPU:** ``` ### 启动 CPU Docker @@ -71,6 +96,9 @@ git clone https://github.com/PaddlePaddle/Serving ### Paddle开发镜像需要执行以下脚本增加Serving所需依赖项 bash Serving/tools/paddle_env_install.sh ``` + + + **GPU:** ``` ### 启动 GPU Docker @@ -103,6 +131,7 @@ pip3 install -r python/requirements.txt ### 2.1 在线安装 +在线安装采用 `pypi` 下载并安装的方式。 ```shell pip3 install paddle-serving-client==0.9.0 -i https://pypi.tuna.tsinghua.edu.cn/simple @@ -158,6 +187,7 @@ pip3 install https://paddle-inference-lib.bj.bcebos.com/2.3.0/python/Linux/GPU/x ### 2.2 离线安装 +离线安装是指所有的 Paddle 和 Serving 包和依赖库,传入到无网或弱网环境下安装。 **1.安装离线 Wheel 包** @@ -223,6 +253,24 @@ python3 install.py --cuda_version="" --python_version="py39" --device="cpu" --se python3 install.py --cuda_version="112" --python_version="py36" --device="GPU" --serving_version="no_install" --paddle_version="2.3.0" ``` + + +### 2.3 ARM & XPU 安装 wheel 包 + +由于使用 ARM 和 XPU 的用户较少,安装此环境的 Wheel 单独提供如下,其中 `paddle_serving_client` 仅提供 `py36` 的版本,如需其他版本请与我们联系。 + +``` +pip3.6 install https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_app-0.9.0-py3-none-any.whl +pip3.6 install https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_client-0.9.0-cp36-none-any.whl +pip3.6 install https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_server_xpu-0.9.0.post2-py3-none-any.whl +``` + +二进制包地址: +``` +wget https://paddle-serving.bj.bcebos.com/test-dev/bin/serving-xpu-aarch64-0.9.0.tar.gz +``` + + ## 3.环境检查 diff --git a/doc/Install_EN.md b/doc/Install_EN.md index 9d6bbf0bf..b62213c14 100644 --- a/doc/Install_EN.md +++ b/doc/Install_EN.md @@ -4,10 +4,16 @@ - [1.Use devel docker](#1) - [Serving devel images](#1.1) + - [CPU images](#1.1.1) + - [GPU images](#1.1.2) + - [ARM & XPU images](#1.1.3) - [Paddle devel images](#1.2) + - [CPU images](#1.2.1) + - [GPU images](#1.2.2) - [2.Install Wheel Packages](#2) - [Online Install](#2.1) - [Offline Install](#2.2) + - [ARM & XPU Install](#2.3) - [3.Installation Check](#3) **Strongly recommend** you build **Paddle Serving** in Docker. For more images, please refer to [Docker Image List](Docker_Images_CN.md). @@ -28,6 +34,7 @@ | CUDA10.2 + cuDNN 7 | 0.9.0-cuda10.2-cudnn7-devel | Ubuntu 16 | 2.3.0-gpu-cuda10.2-cudnn7 | Ubuntu 18 | CUDA10.2 + cuDNN 8 | 0.9.0-cuda10.2-cudnn8-devel | Ubuntu 16 | None | None | | CUDA11.2 + cuDNN 8 | 0.9.0-cuda11.2-cudnn8-devel | Ubuntu 16 | 2.3.0-gpu-cuda11.2-cudnn8 | Ubuntu 18 | +| ARM + XPU | xpu-arm | CentOS 8.3 | None | None | For **Windows 10 users**, please refer to the document [Paddle Serving Guide for Windows Platform](Windows_Tutorial_CN.md). @@ -36,46 +43,68 @@ For **Windows 10 users**, please refer to the document [Paddle Serving Guide for ### 1.1 Serving Devel Images (CPU/GPU 2 choose 1) + + **CPU:** ``` # Start CPU Docker Container docker pull registry.baidubce.com/paddlepaddle/serving:0.9.0-devel -docker run -p 9292:9292 --name test -dit registry.baidubce.com/paddlepaddle/serving:0.9.0-devel bash -docker exec -it test bash +docker run -p 9292:9292 --name test_cpu -dit registry.baidubce.com/paddlepaddle/serving:0.9.0-devel bash +docker exec -it test_cpu bash git clone https://github.com/PaddlePaddle/Serving ``` + + + **GPU:** ``` # Start GPU Docker Container docker pull registry.baidubce.com/paddlepaddle/serving:0.9.0-cuda11.2-cudnn7-devel -nvidia-docker run -p 9292:9292 --name test -dit docker pull registry.baidubce.com/paddlepaddle/serving:0.9.0-cuda11.2-cudnn7-devel bash -nvidia-docker exec -it test bash +nvidia-docker run -p 9292:9292 --name test_gpu -dit docker pull registry.baidubce.com/paddlepaddle/serving:0.9.0-cuda11.2-cudnn7-devel bash +nvidia-docker exec -it test_gpu bash +git clone https://github.com/PaddlePaddle/Serving +``` + + + +**ARM & XPU: ** +``` +docker pull registry.baidubce.com/paddlepaddle/serving:xpu-arm +docker run -p 9292:9292 --name test_arm_xpu -dit registry.baidubce.com/paddlepaddle/serving:xpu-arm bash +docker exec -it test_arm_xpu bash git clone https://github.com/PaddlePaddle/Serving ``` ### 1.2 Paddle Devel Images (choose any codeblock of CPU/GPU) + + + **CPU:** -``` +```shell # Start CPU Docker Container docker pull registry.baidubce.com/paddlepaddle/paddle:2.3.0 -docker run -p 9292:9292 --name test -dit registry.baidubce.com/paddlepaddle/paddle:2.3.0 bash -docker exec -it test bash +docker run -p 9292:9292 --name test_cpu -dit registry.baidubce.com/paddlepaddle/paddle:2.3.0 bash +docker exec -it test_cpu bash git clone https://github.com/PaddlePaddle/Serving -# Paddle dev image needs to run the following script to increase the dependencies required by Serving +### Paddle dev image needs to run the following script to increase the dependencies required by Serving bash Serving/tools/paddle_env_install.sh ``` + + + **GPU:** -``` -# Start GPU Docker + +```shell +### Start GPU Docker nvidia-docker pull registry.baidubce.com/paddlepaddle/paddle:2.3.0-gpu-cuda11.2-cudnn8 -nvidia-docker run -p 9292:9292 --name test -dit registry.baidubce.com/paddlepaddle/paddle:2.3.0-gpu-cuda11.2-cudnn8 bash -nvidia-docker exec -it test bash +nvidia-docker run -p 9292:9292 --name test_gpu -dit registry.baidubce.com/paddlepaddle/paddle:2.3.0-gpu-cuda11.2-cudnn8 bash +nvidia-docker exec -it test_gpu bash git clone https://github.com/PaddlePaddle/Serving -# Paddle development image needs to execute the following script to increase the dependencies required by Serving +### Paddle development image needs to execute the following script to increase the dependencies required by Serving bash Serving/tools/paddle_env_install.sh ``` @@ -98,6 +127,7 @@ Install the service whl package. There are three types of client, app and server ### 2.1 Online Install +Online installation uses `pypi` to download and install. ```shell pip3 install paddle-serving-client==0.9.0 -i https://pypi.tuna.tsinghua.edu.cn/simple @@ -152,6 +182,7 @@ pip3 install https://paddle-inference-lib.bj.bcebos.com/2.3.0/python/Linux/GPU/x ### 2.2 Offline Install +Offline installation is to download all Paddle and Serving packages and dependent libraries, and install them in a no-network or weak-network environment. **1.Install offline wheel packages** @@ -210,6 +241,22 @@ python3 install.py --cuda_version="" --python_version="py39" --device="cpu" --se python3 install.py --cuda_version="112" --python_version="py36" --device="GPU" --serving_version="no_install" --paddle_version="2.3.0" ``` + + +### 2.3 ARM & XPU Install + +Since there are few users using ARM and XPU, the Wheel for installing this environment is provided separately as follows, among which `paddle_serving_client` only provides the `py36` version, if you need other versions, please contact us. +``` +pip3.6 install https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_app-0.9.0-py3-none-any.whl +pip3.6 install https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_client-0.9.0-cp36-none-any.whl +pip3.6 install https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_server_xpu-0.9.0.post2-py3-none-any.whl +``` + +Download binary package address: +``` +wget https://paddle-serving.bj.bcebos.com/test-dev/bin/serving-xpu-aarch64-0.9.0.tar.gz +``` + ## 3.Installation Check diff --git a/doc/Latest_Packages_CN.md b/doc/Latest_Packages_CN.md index ee5a30313..4666b2ea1 100644 --- a/doc/Latest_Packages_CN.md +++ b/doc/Latest_Packages_CN.md @@ -48,22 +48,23 @@ ### Wheel包链接 适用ARM CPU环境的昆仑Wheel包: -``` + +```shell # paddle-serving-server -https://paddle-serving.bj.bcebos.com/whl/xpu/arm/paddle_serving_server_xpu-0.0.0.post2-py3-none-any.whl +wget https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_app-0.9.0-py3-none-any.whl # paddle-serving-client -https://paddle-serving.bj.bcebos.com/whl/xpu/arm/paddle_serving_client-0.0.0-cp36-none-any.whl +wget https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_client-0.9.0-cp36-none-any.whl # paddle-serving-app -https://paddle-serving.bj.bcebos.com/whl/xpu/arm/paddle_serving_app-0.0.0-py3-none-any.whl +wget https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_server_xpu-0.9.0.post2-py3-none-any.whl # SERVING BIN -https://paddle-serving.bj.bcebos.com/bin/serving-xpu-aarch64-0.0.0.tar.gz +wget https://paddle-serving.bj.bcebos.com/test-dev/bin/serving-xpu-aarch64-0.9.0.tar.gz ``` -适用于x86 CPU环境的昆仑Wheel包: -``` +适用于ARM & XPU 的 v0.9.0 版本 Wheel包: +```shell https://paddle-serving.bj.bcebos.com/test-dev/whl/paddle_serving_server_xpu-0.9.0.post2-py3-none-any.whl - ``` + diff --git a/doc/Latest_Packages_EN.md b/doc/Latest_Packages_EN.md index b0cb1397c..3cdcb3449 100644 --- a/doc/Latest_Packages_EN.md +++ b/doc/Latest_Packages_EN.md @@ -46,23 +46,22 @@ for kunlun user who uses arm-xpu or x86-xpu can download the wheel packages as f ### Wheel Package Links -for arm kunlun user -``` +for arm kunlun user, +```shell # paddle-serving-server -https://paddle-serving.bj.bcebos.com/whl/xpu/arm/paddle_serving_server_xpu-0.0.0.post2-py3-none-any.whl +wget https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_app-0.9.0-py3-none-any.whl # paddle-serving-client -https://paddle-serving.bj.bcebos.com/whl/xpu/arm/paddle_serving_client-0.0.0-cp36-none-any.whl +wget https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_client-0.9.0-cp36-none-any.whl # paddle-serving-app -https://paddle-serving.bj.bcebos.com/whl/xpu/arm/paddle_serving_app-0.0.0-py3-none-any.whl +wget https://paddle-serving.bj.bcebos.com/test-dev/whl/arm/paddle_serving_server_xpu-0.9.0.post2-py3-none-any.whl # SERVING BIN -https://paddle-serving.bj.bcebos.com/bin/serving-xpu-aarch64-0.0.0.tar.gz +wget https://paddle-serving.bj.bcebos.com/test-dev/bin/serving-xpu-aarch64-0.9.0.tar.gz ``` -for x86 kunlun user -``` +for x86 xpu user, the wheel packages are here. +```shell https://paddle-serving.bj.bcebos.com/test-dev/whl/paddle_serving_server_xpu-0.9.0.post2-py3-none-any.whl - ```