From 1bf7606e0beb7c62ee97a5f8f3ef15af1d35c3f7 Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Fri, 15 Aug 2025 16:14:51 +0200 Subject: [PATCH] chore(rocm): bump rocm image, add gfx1200 support Fixes: https://github.com/mudler/LocalAI/issues/6044 Signed-off-by: Ettore Di Giacinto --- .github/workflows/backend.yml | 24 ++++++++++++------------ .github/workflows/image-pr.yml | 2 +- .github/workflows/image.yml | 2 +- backend/cpp/llama-cpp/Makefile | 2 +- 4 files changed, 15 insertions(+), 15 deletions(-) diff --git a/.github/workflows/backend.yml b/.github/workflows/backend.yml index 926ebc521437..3ff701d76864 100644 --- a/.github/workflows/backend.yml +++ b/.github/workflows/backend.yml @@ -290,7 +290,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-rerankers' runs-on: 'ubuntu-latest' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "rerankers" dockerfile: "./backend/Dockerfile.python" @@ -302,7 +302,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-llama-cpp' runs-on: 'ubuntu-latest' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "llama-cpp" dockerfile: "./backend/Dockerfile.llama-cpp" @@ -314,7 +314,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-vllm' runs-on: 'ubuntu-latest' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "vllm" dockerfile: "./backend/Dockerfile.python" @@ -326,7 +326,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-transformers' runs-on: 'arc-runner-set' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "transformers" dockerfile: "./backend/Dockerfile.python" @@ -338,7 +338,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-diffusers' runs-on: 'arc-runner-set' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "diffusers" dockerfile: "./backend/Dockerfile.python" @@ -351,7 +351,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-kokoro' runs-on: 'arc-runner-set' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "kokoro" dockerfile: "./backend/Dockerfile.python" @@ -363,7 +363,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-faster-whisper' runs-on: 'ubuntu-latest' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "faster-whisper" dockerfile: "./backend/Dockerfile.python" @@ -375,7 +375,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-coqui' runs-on: 'ubuntu-latest' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "coqui" dockerfile: "./backend/Dockerfile.python" @@ -387,7 +387,7 @@ jobs: tag-latest: 'auto' tag-suffix: '-gpu-rocm-hipblas-bark' runs-on: 'arc-runner-set' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' backend: "bark" dockerfile: "./backend/Dockerfile.python" @@ -752,7 +752,7 @@ jobs: platforms: 'linux/amd64' tag-latest: 'auto' tag-suffix: '-gpu-hipblas-whisper' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" runs-on: 'ubuntu-latest' skip-drivers: 'false' backend: "whisper" @@ -914,7 +914,7 @@ jobs: skip-drivers: 'true' tag-latest: 'auto' tag-suffix: '-gpu-hipblas-exllama2' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" runs-on: 'ubuntu-latest' backend: "exllama2" dockerfile: "./backend/Dockerfile.python" @@ -926,7 +926,7 @@ jobs: # platforms: 'linux/amd64' # tag-latest: 'auto' # tag-suffix: '-gpu-hipblas-rfdetr' - # base-image: "rocm/dev-ubuntu-22.04:6.1" + # base-image: "rocm/dev-ubuntu-22.04:6.4.3" # runs-on: 'ubuntu-latest' # skip-drivers: 'false' # backend: "rfdetr" diff --git a/.github/workflows/image-pr.yml b/.github/workflows/image-pr.yml index c6577be8b279..a2410b22827e 100644 --- a/.github/workflows/image-pr.yml +++ b/.github/workflows/image-pr.yml @@ -47,7 +47,7 @@ jobs: platforms: 'linux/amd64' tag-latest: 'false' tag-suffix: '-hipblas' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" grpc-base-image: "ubuntu:22.04" runs-on: 'ubuntu-latest' makeflags: "--jobs=3 --output-sync=target" diff --git a/.github/workflows/image.yml b/.github/workflows/image.yml index 3055a18d7be0..3864930d03ed 100644 --- a/.github/workflows/image.yml +++ b/.github/workflows/image.yml @@ -39,7 +39,7 @@ jobs: platforms: 'linux/amd64' tag-latest: 'auto' tag-suffix: '-gpu-hipblas' - base-image: "rocm/dev-ubuntu-22.04:6.1" + base-image: "rocm/dev-ubuntu-22.04:6.4.3" grpc-base-image: "ubuntu:22.04" runs-on: 'ubuntu-latest' makeflags: "--jobs=3 --output-sync=target" diff --git a/backend/cpp/llama-cpp/Makefile b/backend/cpp/llama-cpp/Makefile index af72f3635854..a2bc04c8c1cd 100644 --- a/backend/cpp/llama-cpp/Makefile +++ b/backend/cpp/llama-cpp/Makefile @@ -32,7 +32,7 @@ else ifeq ($(BUILD_TYPE),hipblas) ROCM_PATH ?= /opt/rocm export CXX=$(ROCM_HOME)/llvm/bin/clang++ export CC=$(ROCM_HOME)/llvm/bin/clang - AMDGPU_TARGETS?=gfx803,gfx900,gfx906,gfx908,gfx90a,gfx942,gfx1010,gfx1030,gfx1032,gfx1100,gfx1101,gfx1102 + AMDGPU_TARGETS?=gfx803,gfx900,gfx906,gfx908,gfx90a,gfx942,gfx1010,gfx1030,gfx1032,gfx1100,gfx1101,gfx1102,gfx1200 CMAKE_ARGS+=-DGGML_HIP=ON -DAMDGPU_TARGETS=$(AMDGPU_TARGETS) else ifeq ($(BUILD_TYPE),vulkan) CMAKE_ARGS+=-DGGML_VULKAN=1