From ebd64ce12ce402a7e69135cbb164d1cf10cc58c2 Mon Sep 17 00:00:00 2001 From: Kibae Shin Date: Sun, 17 Sep 2023 21:57:55 +0900 Subject: [PATCH] ci: docker 1.2.1 release (#34) --- README.md | 6 +++--- deploy/build-docker/VERSION | 2 +- deploy/build-docker/docker-compose.yaml | 4 ++-- docs/docker.md | 10 +++++----- docs/swagger/openapi.yaml | 2 +- 5 files changed, 12 insertions(+), 12 deletions(-) diff --git a/README.md b/README.md index 8e7d096..dea2720 100644 --- a/README.md +++ b/README.md @@ -147,13 +147,13 @@ sudo cmake --install build --prefix /usr/local/onnxruntime-server # Docker - Docker hub: [kibaes/onnxruntime-server](https://hub.docker.com/r/kibaes/onnxruntime-server) - - [`1.2.0-linux-cuda`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cuda.dockerfile) + - [`1.2.1-linux-cuda`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cuda.dockerfile) amd64 - - [`1.2.0-linux-cpu`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cpu.dockerfile) + - [`1.2.1-linux-cpu`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cpu.dockerfile) amd64, arm64 ```shell -DOCKER_IMAGE=kibae/onnxruntime-server:1.2.0-linux-cuda # or kibae/onnxruntime-server:1.2.0-linux-cpu +DOCKER_IMAGE=kibae/onnxruntime-server:1.2.1-linux-cuda # or kibae/onnxruntime-server:1.2.1-linux-cpu docker pull ${DOCKER_IMAGE} diff --git a/deploy/build-docker/VERSION b/deploy/build-docker/VERSION index fa31d2b..9b81b9d 100644 --- a/deploy/build-docker/VERSION +++ b/deploy/build-docker/VERSION @@ -1,2 +1,2 @@ -export VERSION=1.2.0 +export VERSION=1.2.1 export IMAGE_PREFIX=kibaes/onnxruntime-server diff --git a/deploy/build-docker/docker-compose.yaml b/deploy/build-docker/docker-compose.yaml index b8a1392..2ae2fd7 100644 --- a/deploy/build-docker/docker-compose.yaml +++ b/deploy/build-docker/docker-compose.yaml @@ -5,7 +5,7 @@ services: onnxruntime_server_simple: # After the docker container is up, you can use the REST API (http://localhost:8080). # API documentation will be available at http://localhost:8080/api-docs. - image: kibaes/onnxruntime-server:1.2.0-linux-cuda + image: kibaes/onnxruntime-server:1.2.1-linux-cuda ports: - "8080:80" # for http backend volumes: @@ -29,7 +29,7 @@ services: onnxruntime_server_advanced: # After the docker container is up, you can use the REST API (http://localhost, https://localhost). # API documentation will be available at http://localhost/api-docs. - image: kibaes/onnxruntime-server:1.2.0-linux-cuda + image: kibaes/onnxruntime-server:1.2.1-linux-cuda ports: - "80:80" # for http backend - "443:443" # for https backend diff --git a/docs/docker.md b/docs/docker.md index 335b402..e4ff419 100644 --- a/docs/docker.md +++ b/docs/docker.md @@ -5,9 +5,9 @@ # Supported tags and respective Dockerfile links -- [`1.2.0-linux-cuda`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cuda.dockerfile) +- [`1.2.1-linux-cuda`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cuda.dockerfile) amd64 -- [`1.2.0-linux-cpu`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cpu.dockerfile) +- [`1.2.1-linux-cpu`](https://github.com/kibae/onnxruntime-server/blob/main/deploy/build-docker/linux-cpu.dockerfile) amd64, arm64 # How to use this image @@ -30,7 +30,7 @@ - API documentation will be available at http://localhost/api-docs. ```shell -DOCKER_IMAGE=kibae/onnxruntime-server:1.2.0-linux-cuda # or kibae/onnxruntime-server:1.2.0-linux-cpu +DOCKER_IMAGE=kibae/onnxruntime-server:1.2.1-linux-cuda # or kibae/onnxruntime-server:1.2.1-linux-cpu docker pull ${DOCKER_IMAGE} @@ -71,7 +71,7 @@ services: onnxruntime_server_simple: # After the docker container is up, you can use the REST API (http://localhost:8080). # API documentation will be available at http://localhost:8080/api-docs. - image: kibaes/onnxruntime-server:1.2.0-linux-cuda + image: kibaes/onnxruntime-server:1.2.1-linux-cuda ports: - "8080:80" # for http backend volumes: @@ -103,7 +103,7 @@ services: onnxruntime_server_advanced: # After the docker container is up, you can use the REST API (http://localhost, https://localhost). # API documentation will be available at http://localhost/api-docs. - image: kibaes/onnxruntime-server:1.2.0-linux-cuda + image: kibaes/onnxruntime-server:1.2.1-linux-cuda ports: - "80:80" # for http backend - "443:443" # for https backend diff --git a/docs/swagger/openapi.yaml b/docs/swagger/openapi.yaml index 7111cac..fa2fd39 100644 --- a/docs/swagger/openapi.yaml +++ b/docs/swagger/openapi.yaml @@ -2,7 +2,7 @@ openapi: 3.0.3 info: title: ONNX Runtime Server description: |- - version: 1.2.0 + version: 1.2.1 externalDocs: description: ONNX Runtime Server url: https://github.com/kibae/onnxruntime-server