From 5f52a10ffef342ef7ab84e9cf7107903d1e578e4 Mon Sep 17 00:00:00 2001 From: "chen, suyue" Date: Mon, 8 Jul 2024 09:05:27 +0800 Subject: [PATCH] update tgi version (#378) * try with tgi-gaudi:2.0.0 Signed-off-by: chensuyue * try with text-generation-inference:2.1.0 Signed-off-by: chensuyue * add params for tgi-gaudi:2.0.0 Signed-off-by: chensuyue --------- Signed-off-by: chensuyue --- ChatQnA/docker/gaudi/docker_compose.yaml | 4 ++-- ChatQnA/docker/xeon/docker_compose.yaml | 2 +- ChatQnA/tests/test_chatqna_on_gaudi.sh | 2 +- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/ChatQnA/docker/gaudi/docker_compose.yaml b/ChatQnA/docker/gaudi/docker_compose.yaml index 0e709c449..d297d4e9f 100644 --- a/ChatQnA/docker/gaudi/docker_compose.yaml +++ b/ChatQnA/docker/gaudi/docker_compose.yaml @@ -114,7 +114,7 @@ services: LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: - image: ghcr.io/huggingface/tgi-gaudi:1.2.1 + image: ghcr.io/huggingface/tgi-gaudi:2.0.0 container_name: tgi-gaudi-server ports: - "8008:80" @@ -133,7 +133,7 @@ services: cap_add: - SYS_NICE ipc: host - command: --model-id ${LLM_MODEL_ID} + command: --model-id ${LLM_MODEL_ID} --max-input-length 1024 --max-total-tokens 2048 llm: image: opea/llm-tgi:latest container_name: llm-tgi-gaudi-server diff --git a/ChatQnA/docker/xeon/docker_compose.yaml b/ChatQnA/docker/xeon/docker_compose.yaml index 9cc80c82e..55ce080ed 100644 --- a/ChatQnA/docker/xeon/docker_compose.yaml +++ b/ChatQnA/docker/xeon/docker_compose.yaml @@ -110,7 +110,7 @@ services: LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi_service: - image: ghcr.io/huggingface/text-generation-inference:1.4 + image: ghcr.io/huggingface/text-generation-inference:2.1.0 container_name: tgi-service ports: - "9009:80" diff --git a/ChatQnA/tests/test_chatqna_on_gaudi.sh b/ChatQnA/tests/test_chatqna_on_gaudi.sh index 4055de051..f4abc961e 100644 --- a/ChatQnA/tests/test_chatqna_on_gaudi.sh +++ b/ChatQnA/tests/test_chatqna_on_gaudi.sh @@ -25,7 +25,7 @@ function build_docker_images() { cd tei-gaudi/ docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest . - docker pull ghcr.io/huggingface/tgi-gaudi:1.2.1 + docker pull ghcr.io/huggingface/tgi-gaudi:2.0.0 docker pull ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 cd $WORKPATH/docker