diff --git a/.github/workflows/_gmc-e2e.yml b/.github/workflows/_gmc-e2e.yml index ba50e8b955..debd59ea03 100644 --- a/.github/workflows/_gmc-e2e.yml +++ b/.github/workflows/_gmc-e2e.yml @@ -56,7 +56,7 @@ jobs: - name: Run tests id: run-test env: - HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} + HF_TOKEN: ${{ secrets.HF_TOKEN }} GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }} GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }} run: | diff --git a/.github/workflows/_helm-e2e.yml b/.github/workflows/_helm-e2e.yml index 61a91ce583..fa7ca06761 100644 --- a/.github/workflows/_helm-e2e.yml +++ b/.github/workflows/_helm-e2e.yml @@ -165,8 +165,8 @@ jobs: env: GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }} GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }} - HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} - HFTOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} + HF_TOKEN: ${{ secrets.HF_TOKEN }} + HFTOKEN: ${{ secrets.HF_TOKEN }} value_file: ${{ matrix.value_file }} run: | set -xe @@ -190,7 +190,7 @@ jobs: --namespace $NAMESPACE \ $RELEASE_NAME \ oci://ghcr.io/opea-project/charts/${CHART_NAME} \ - --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} \ + --set global.HF_TOKEN=${HFTOKEN} \ --set global.modelUseHostPath=/data2/hf_model \ --set GOOGLE_API_KEY=${{ env.GOOGLE_API_KEY}} \ --set GOOGLE_CSE_ID=${{ env.GOOGLE_CSE_ID}} \ diff --git a/.github/workflows/_run-docker-compose.yml b/.github/workflows/_run-docker-compose.yml index 7af15e11be..4301de9ebc 100644 --- a/.github/workflows/_run-docker-compose.yml +++ b/.github/workflows/_run-docker-compose.yml @@ -170,8 +170,8 @@ jobs: - name: Run test shell: bash env: - HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} - HF_TOKEN: ${{ secrets.HUGGINGFACEHUB_API_TOKEN }} + HUGGINGFACEHUB_API_TOKEN: ${{ secrets.HF_TOKEN }} + HF_TOKEN: ${{ secrets.HF_TOKEN }} GOOGLE_CSE_ID: ${{ secrets.GOOGLE_CSE_ID }} GOOGLE_API_KEY: ${{ secrets.GOOGLE_API_KEY }} PINECONE_KEY: ${{ secrets.PINECONE_KEY }} diff --git a/AgentQnA/README.md b/AgentQnA/README.md index c78703d6fb..e7f01a3ee1 100644 --- a/AgentQnA/README.md +++ b/AgentQnA/README.md @@ -123,7 +123,7 @@ Set up a [HuggingFace](https://huggingface.co/) account and generate a [user acc Then set an environment variable with the token and another for a directory to download the models: ```bash -export HUGGINGFACEHUB_API_TOKEN= +export HF_TOKEN= export HF_CACHE_DIR= # to avoid redownloading models ``` diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/README.md b/AgentQnA/docker_compose/amd/gpu/rocm/README.md index fe5253ed07..27854510e5 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/README.md +++ b/AgentQnA/docker_compose/amd/gpu/rocm/README.md @@ -170,7 +170,7 @@ Use AMD GPU driver utilities to determine the correct `cardN` and `renderN` IDs ### Replace the string 'server_address' with your local server IP address export host_ip='server_address' ### Replace the string 'your_huggingfacehub_token' with your HuggingFacehub repository access token. -export HUGGINGFACEHUB_API_TOKEN='your_huggingfacehub_token' +export HF_TOKEN='your_huggingfacehub_token' ### Replace the string 'your_langchain_api_key' with your LANGCHAIN API KEY. export LANGCHAIN_API_KEY='your_langchain_api_key' export LANGCHAIN_TRACING_V2="" diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/compose.yaml b/AgentQnA/docker_compose/amd/gpu/rocm/compose.yaml index 4eab372dec..722019c6e0 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/compose.yaml +++ b/AgentQnA/docker_compose/amd/gpu/rocm/compose.yaml @@ -13,8 +13,8 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: "http://${ip_address}:${TGI_SERVICE_PORT}" - HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} shm_size: 32g devices: - /dev/kfd:/dev/kfd @@ -42,7 +42,7 @@ services: with_memory: false recursion_limit: ${recursion_limit_worker} llm_engine: tgi - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -76,7 +76,7 @@ services: use_hints: false recursion_limit: ${recursion_limit_worker} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -104,7 +104,7 @@ services: with_memory: true recursion_limit: ${recursion_limit_supervisor} llm_engine: tgi - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/compose_vllm.yaml b/AgentQnA/docker_compose/amd/gpu/rocm/compose_vllm.yaml index 4b37d15d10..28e48fc569 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/compose_vllm.yaml +++ b/AgentQnA/docker_compose/amd/gpu/rocm/compose_vllm.yaml @@ -10,8 +10,8 @@ services: no_proxy: ${no_proxy} http_proxy: ${http_proxy} https_proxy: ${https_proxy} - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} + HF_TOKEN: ${HF_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 WILM_USE_TRITON_FLASH_ATTENTION: 0 @@ -46,7 +46,7 @@ services: with_memory: false recursion_limit: ${recursion_limit_worker} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -80,7 +80,7 @@ services: use_hints: false recursion_limit: ${recursion_limit_worker} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -108,7 +108,7 @@ services: with_memory: true recursion_limit: ${recursion_limit_supervisor} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_tgi_rocm.sh b/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_tgi_rocm.sh index 209d9b372a..7774202208 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_tgi_rocm.sh +++ b/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_tgi_rocm.sh @@ -1,9 +1,10 @@ +#!/bin/bash # Copyright (C) 2024 Advanced Micro Devices, Inc. # SPDX-License-Identifier: Apache-2.0 # Before start script: # export host_ip="your_host_ip_or_host_name" -# export HUGGINGFACEHUB_API_TOKEN="your_huggingface_api_token" +# export HF_TOKEN="your_huggingface_api_token" # export LANGCHAIN_API_KEY="your_langchain_api_key" # export LANGCHAIN_TRACING_V2="" @@ -19,7 +20,7 @@ export CRAG_SERVER_PORT="18114" export WORKPATH=$(dirname "$PWD") export WORKDIR=${WORKPATH}/../../../ -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export HF_CACHE_DIR="./data" export MODEL_CACHE="./data" @@ -39,7 +40,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT} export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions" export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions" export HF_CACHE_DIR=${HF_CACHE_DIR} -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export no_proxy=${no_proxy} export http_proxy=${http_proxy} export https_proxy=${https_proxy} diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_vllm_rocm.sh b/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_vllm_rocm.sh index 2d28ea3c34..1e7a7289bd 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_vllm_rocm.sh +++ b/AgentQnA/docker_compose/amd/gpu/rocm/launch_agent_service_vllm_rocm.sh @@ -1,9 +1,10 @@ +#!/bin/bash # Copyright (C) 2024 Advanced Micro Devices, Inc. # SPDX-License-Identifier: Apache-2.0 # Before start script: # export host_ip="your_host_ip_or_host_name" -# export HUGGINGFACEHUB_API_TOKEN="your_huggingface_api_token" +# export HF_TOKEN="your_huggingface_api_token" # export LANGCHAIN_API_KEY="your_langchain_api_key" # export LANGCHAIN_TRACING_V2="" @@ -19,7 +20,7 @@ export CRAG_SERVER_PORT="18114" export WORKPATH=$(dirname "$PWD") export WORKDIR=${WORKPATH}/../../../ -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export HF_CACHE_DIR="./data" export MODEL_CACHE="./data" @@ -40,7 +41,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT} export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions" export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions" export HF_CACHE_DIR=${HF_CACHE_DIR} -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export no_proxy=${no_proxy} export http_proxy=${http_proxy} export https_proxy=${https_proxy} diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_tgi_rocm.sh b/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_tgi_rocm.sh index ab5b133c18..58c6c303cb 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_tgi_rocm.sh +++ b/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_tgi_rocm.sh @@ -5,7 +5,7 @@ WORKPATH=$(dirname "$PWD")/.. export ip_address=${host_ip} -export HUGGINGFACEHUB_API_TOKEN=${your_hf_api_token} +export HF_TOKEN=${HF_TOKEN} export AGENTQNA_TGI_IMAGE=ghcr.io/huggingface/text-generation-inference:2.4.1-rocm export AGENTQNA_TGI_SERVICE_PORT="19001" diff --git a/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_vllm_rocm.sh b/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_vllm_rocm.sh index d448864e08..fa9d6508e6 100644 --- a/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_vllm_rocm.sh +++ b/AgentQnA/docker_compose/amd/gpu/rocm/stop_agent_service_vllm_rocm.sh @@ -1,10 +1,11 @@ +#!/bin/bash # Copyright (C) 2024 Advanced Micro Devices, Inc. # SPDX-License-Identifier: Apache-2.0 # Before start script: # export host_ip="your_host_ip_or_host_name" -# export HUGGINGFACEHUB_API_TOKEN="your_huggingface_api_token" +# export HF_TOKEN="your_huggingface_api_token" # export LANGCHAIN_API_KEY="your_langchain_api_key" # export LANGCHAIN_TRACING_V2="" @@ -20,8 +21,7 @@ export CRAG_SERVER_PORT="18114" export WORKPATH=$(dirname "$PWD") export WORKDIR=${WORKPATH}/../../../ -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export VLLM_LLM_MODEL_ID="Intel/neural-chat-7b-v3-3" export HF_CACHE_DIR="./data" export MODEL_CACHE="./data" @@ -42,7 +42,7 @@ export CRAG_SERVER=http://${ip_address}:${CRAG_SERVER_PORT} export WORKER_AGENT_URL="http://${ip_address}:${WORKER_RAG_AGENT_PORT}/v1/chat/completions" export SQL_AGENT_URL="http://${ip_address}:${WORKER_SQL_AGENT_PORT}/v1/chat/completions" export HF_CACHE_DIR=${HF_CACHE_DIR} -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export no_proxy=${no_proxy} export http_proxy=${http_proxy} export https_proxy=${https_proxy} diff --git a/AgentQnA/docker_compose/intel/cpu/xeon/set_env.sh b/AgentQnA/docker_compose/intel/cpu/xeon/set_env.sh index ba452cf4bb..470cfb044e 100644 --- a/AgentQnA/docker_compose/intel/cpu/xeon/set_env.sh +++ b/AgentQnA/docker_compose/intel/cpu/xeon/set_env.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 @@ -33,7 +34,7 @@ fi # retriever export host_ip=$(hostname -I | awk '{print $1}') export HF_CACHE_DIR=${HF_CACHE_DIR} -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export no_proxy=${no_proxy} export http_proxy=${http_proxy} export https_proxy=${https_proxy} diff --git a/AgentQnA/docker_compose/intel/hpu/gaudi/compose.yaml b/AgentQnA/docker_compose/intel/hpu/gaudi/compose.yaml index a5240ad4b8..9945fa2b41 100644 --- a/AgentQnA/docker_compose/intel/hpu/gaudi/compose.yaml +++ b/AgentQnA/docker_compose/intel/hpu/gaudi/compose.yaml @@ -16,7 +16,7 @@ services: with_memory: false recursion_limit: ${recursion_limit_worker} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -50,7 +50,7 @@ services: use_hints: false recursion_limit: ${recursion_limit_worker} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -79,7 +79,7 @@ services: with_memory: true recursion_limit: ${recursion_limit_supervisor} llm_engine: vllm - HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGINGFACEHUB_API_TOKEN: ${HF_TOKEN} llm_endpoint_url: ${LLM_ENDPOINT_URL} model: ${LLM_MODEL_ID} temperature: ${temperature} @@ -122,7 +122,7 @@ services: no_proxy: ${no_proxy} http_proxy: ${http_proxy} https_proxy: ${https_proxy} - HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HF_TOKEN: ${HF_TOKEN} HABANA_VISIBLE_DEVICES: all OMPI_MCA_btl_vader_single_copy_mechanism: none LLM_MODEL_ID: ${LLM_MODEL_ID} diff --git a/AgentQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh b/AgentQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh index df4012d2d3..a37e6c39d8 100644 --- a/AgentQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh +++ b/AgentQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2025 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/AgentQnA/docker_compose/intel/hpu/gaudi/set_env.sh b/AgentQnA/docker_compose/intel/hpu/gaudi/set_env.sh index cc8139c960..fb66c459e3 100644 --- a/AgentQnA/docker_compose/intel/hpu/gaudi/set_env.sh +++ b/AgentQnA/docker_compose/intel/hpu/gaudi/set_env.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 @@ -16,8 +17,7 @@ export ip_address=$(hostname -I | awk '{print $1}') # LLM related environment variables export HF_CACHE_DIR=${HF_CACHE_DIR} ls $HF_CACHE_DIR -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -export HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export LLM_MODEL_ID="meta-llama/Llama-3.3-70B-Instruct" export NUM_SHARDS=4 export LLM_ENDPOINT_URL="http://${ip_address}:8086" diff --git a/AgentQnA/docker_compose/intel/hpu/gaudi/tgi_gaudi.yaml b/AgentQnA/docker_compose/intel/hpu/gaudi/tgi_gaudi.yaml index 856642456a..86324b8652 100644 --- a/AgentQnA/docker_compose/intel/hpu/gaudi/tgi_gaudi.yaml +++ b/AgentQnA/docker_compose/intel/hpu/gaudi/tgi_gaudi.yaml @@ -13,7 +13,7 @@ services: no_proxy: ${no_proxy} http_proxy: ${http_proxy} https_proxy: ${https_proxy} - HUGGING_FACE_HUB_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} + HUGGING_FACE_HUB_TOKEN: ${HF_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 HABANA_VISIBLE_DEVICES: all diff --git a/AgentQnA/retrieval_tool/launch_retrieval_tool.sh b/AgentQnA/retrieval_tool/launch_retrieval_tool.sh index 40ff1ff62a..b5d1a76686 100644 --- a/AgentQnA/retrieval_tool/launch_retrieval_tool.sh +++ b/AgentQnA/retrieval_tool/launch_retrieval_tool.sh @@ -1,9 +1,10 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 host_ip=$(hostname -I | awk '{print $1}') export HF_CACHE_DIR=${HF_CACHE_DIR} -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export no_proxy=${no_proxy} export http_proxy=${http_proxy} export https_proxy=${https_proxy} diff --git a/AgentQnA/retrieval_tool/run_ingest_data.sh b/AgentQnA/retrieval_tool/run_ingest_data.sh index 8dcd5a22fc..00a3c4e785 100644 --- a/AgentQnA/retrieval_tool/run_ingest_data.sh +++ b/AgentQnA/retrieval_tool/run_ingest_data.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/AgentQnA/tests/sql_agent_test/run_data_split.sh b/AgentQnA/tests/sql_agent_test/run_data_split.sh index 2fc2dfcb0e..07f1d71474 100644 --- a/AgentQnA/tests/sql_agent_test/run_data_split.sh +++ b/AgentQnA/tests/sql_agent_test/run_data_split.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/AgentQnA/tests/step2_start_retrieval_tool_rocm_vllm.sh b/AgentQnA/tests/step2_start_retrieval_tool_rocm_vllm.sh index 1c039ad6a0..6a15e79c37 100644 --- a/AgentQnA/tests/step2_start_retrieval_tool_rocm_vllm.sh +++ b/AgentQnA/tests/step2_start_retrieval_tool_rocm_vllm.sh @@ -20,7 +20,7 @@ function start_retrieval_tool() { cd $WORKPATH/../DocIndexRetriever/docker_compose/intel/cpu/xeon host_ip=$(hostname -I | awk '{print $1}') export HF_CACHE_DIR=${HF_CACHE_DIR} - export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} + export HF_TOKEN=${HF_TOKEN} export no_proxy=${no_proxy} export http_proxy=${http_proxy} export https_proxy=${https_proxy} diff --git a/AgentQnA/tests/step4_launch_and_validate_agent_gaudi.sh b/AgentQnA/tests/step4_launch_and_validate_agent_gaudi.sh index 6e970ce2e8..b853094c50 100644 --- a/AgentQnA/tests/step4_launch_and_validate_agent_gaudi.sh +++ b/AgentQnA/tests/step4_launch_and_validate_agent_gaudi.sh @@ -11,8 +11,7 @@ export ip_address=$(hostname -I | awk '{print $1}') export host_ip=$ip_address echo "ip_address=${ip_address}" export TOOLSET_PATH=$WORKPATH/tools/ -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} -HF_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} model="meta-llama/Llama-3.3-70B-Instruct" #"meta-llama/Meta-Llama-3.1-70B-Instruct" export HF_CACHE_DIR=${model_cache:-"/data2/huggingface"} diff --git a/AgentQnA/tests/test_compose_on_gaudi.sh b/AgentQnA/tests/test_compose_on_gaudi.sh index f6e7e95997..2920455e2b 100644 --- a/AgentQnA/tests/test_compose_on_gaudi.sh +++ b/AgentQnA/tests/test_compose_on_gaudi.sh @@ -7,7 +7,7 @@ WORKPATH=$(dirname "$PWD") export WORKDIR=$WORKPATH/../../ echo "WORKDIR=${WORKDIR}" export ip_address=$(hostname -I | awk '{print $1}') -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export TOOLSET_PATH=$WORKDIR/GenAIExamples/AgentQnA/tools/ export no_proxy="$no_proxy,rag-agent-endpoint,sql-agent-endpoint,react-agent-endpoint,agent-ui,vllm-gaudi-server,jaeger,grafana,prometheus,127.0.0.1,localhost,0.0.0.0,$ip_address" IMAGE_REPO=${IMAGE_REPO:-"opea"} diff --git a/AgentQnA/tests/test_compose_on_rocm.sh b/AgentQnA/tests/test_compose_on_rocm.sh index 561ab0a967..1a466fff72 100644 --- a/AgentQnA/tests/test_compose_on_rocm.sh +++ b/AgentQnA/tests/test_compose_on_rocm.sh @@ -9,7 +9,7 @@ ls $WORKPATH export WORKDIR=$WORKPATH/../../ echo "WORKDIR=${WORKDIR}" export ip_address=$(hostname -I | awk '{print $1}') -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export TOOLSET_PATH=$WORKPATH/tools/ IMAGE_REPO=${IMAGE_REPO:-"opea"} IMAGE_TAG=${IMAGE_TAG:-"latest"} diff --git a/AgentQnA/tests/test_compose_vllm_on_rocm.sh b/AgentQnA/tests/test_compose_vllm_on_rocm.sh index 711554f965..01630ee243 100644 --- a/AgentQnA/tests/test_compose_vllm_on_rocm.sh +++ b/AgentQnA/tests/test_compose_vllm_on_rocm.sh @@ -8,7 +8,7 @@ WORKPATH=$(dirname "$PWD") export WORKDIR=${WORKPATH}/../../ echo "WORKDIR=${WORKDIR}" export ip_address=$(hostname -I | awk '{print $1}') -export HUGGINGFACEHUB_API_TOKEN=${HUGGINGFACEHUB_API_TOKEN} +export HF_TOKEN=${HF_TOKEN} export TOOLSET_PATH=$WORKPATH/tools/ IMAGE_REPO=${IMAGE_REPO:-"opea"} IMAGE_TAG=${IMAGE_TAG:-"latest"} diff --git a/ChatQnA/benchmark/accuracy_faqgen/launch_tgi.sh b/ChatQnA/benchmark/accuracy_faqgen/launch_tgi.sh index 72082d1419..0af4431184 100644 --- a/ChatQnA/benchmark/accuracy_faqgen/launch_tgi.sh +++ b/ChatQnA/benchmark/accuracy_faqgen/launch_tgi.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/ChatQnA/benchmark/accuracy_faqgen/run_acc.sh b/ChatQnA/benchmark/accuracy_faqgen/run_acc.sh index 766b718ff8..d31cc74080 100644 --- a/ChatQnA/benchmark/accuracy_faqgen/run_acc.sh +++ b/ChatQnA/benchmark/accuracy_faqgen/run_acc.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/ChatQnA/docker_compose/intel/cpu/xeon/grafana/dashboards/download_opea_dashboard.sh b/ChatQnA/docker_compose/intel/cpu/xeon/grafana/dashboards/download_opea_dashboard.sh index c3739ec705..b98476de2a 100644 --- a/ChatQnA/docker_compose/intel/cpu/xeon/grafana/dashboards/download_opea_dashboard.sh +++ b/ChatQnA/docker_compose/intel/cpu/xeon/grafana/dashboards/download_opea_dashboard.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2025 Intel Corporation # SPDX-License-Identifier: Apache-2.0 if ls *.json 1> /dev/null 2>&1; then diff --git a/ChatQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh b/ChatQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh index 03e0d1f3c2..598ed1427a 100644 --- a/ChatQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh +++ b/ChatQnA/docker_compose/intel/hpu/gaudi/grafana/dashboards/download_opea_dashboard.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2025 Intel Corporation # SPDX-License-Identifier: Apache-2.0 rm *.json diff --git a/ChatQnA/docker_compose/intel/hpu/gaudi/set_env.sh b/ChatQnA/docker_compose/intel/hpu/gaudi/set_env.sh index fe847e6036..1d0409eccd 100755 --- a/ChatQnA/docker_compose/intel/hpu/gaudi/set_env.sh +++ b/ChatQnA/docker_compose/intel/hpu/gaudi/set_env.sh @@ -1,4 +1,4 @@ -#/usr/bin/env bash +#!/usr/bin/env bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/ChatQnA/entrypoint.sh b/ChatQnA/entrypoint.sh index ae395cde0e..c45ae5f94b 100644 --- a/ChatQnA/entrypoint.sh +++ b/ChatQnA/entrypoint.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2025 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/HybridRAG/tests/test_compose_on_gaudi.sh b/HybridRAG/tests/test_compose_on_gaudi.sh index f8c2ccf203..9c512afc9a 100755 --- a/HybridRAG/tests/test_compose_on_gaudi.sh +++ b/HybridRAG/tests/test_compose_on_gaudi.sh @@ -99,7 +99,7 @@ function validate_service() { local HTTP_STATUS=$(curl -s -o /dev/null -w "%{http_code}" -X POST -d "$INPUT_DATA" -H 'Content-Type: application/json' "$URL") - if [ "DOCKER_NAME" -eq "text2cypher-gaudi-container" ]; then + if [ "$DOCKER_NAME" == "text2cypher-gaudi-container" ]; then docker ps docker logs text2cypher-gaudi-container fi @@ -114,7 +114,7 @@ function validate_service() { else echo "[ $SERVICE_NAME ] Content does not match the expected result: $CONTENT" docker logs ${DOCKER_NAME} >> ${LOG_PATH}/${SERVICE_NAME}.log - if [ "DOCKER_NAME" -eq "hybridrag-xeon-backend-server" ]; then + if [ "$DOCKER_NAME" == "hybridrag-xeon-backend-server" ]; then docker ps docker logs text2cypher-gaudi-container fi @@ -123,7 +123,7 @@ function validate_service() { else echo "[ $SERVICE_NAME ] HTTP status is not 200. Received status was $HTTP_STATUS" docker logs ${DOCKER_NAME} >> ${LOG_PATH}/${SERVICE_NAME}.log - if [ "DOCKER_NAME" -eq "hybridrag-xeon-backend-server" ]; then + if [ "$DOCKER_NAME" == "hybridrag-xeon-backend-server" ]; then docker ps docker logs text2cypher-gaudi-container fi diff --git a/Text2Image/tests/test_compose_on_gaudi.sh b/Text2Image/tests/test_compose_on_gaudi.sh index 2e4e70c84d..811782cd5b 100644 --- a/Text2Image/tests/test_compose_on_gaudi.sh +++ b/Text2Image/tests/test_compose_on_gaudi.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0 diff --git a/Text2Image/tests/test_compose_on_xeon.sh b/Text2Image/tests/test_compose_on_xeon.sh index ac2f344482..650c5d47ed 100644 --- a/Text2Image/tests/test_compose_on_xeon.sh +++ b/Text2Image/tests/test_compose_on_xeon.sh @@ -1,3 +1,4 @@ +#!/bin/bash # Copyright (C) 2024 Intel Corporation # SPDX-License-Identifier: Apache-2.0