Skip to content
Open
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
273 changes: 273 additions & 0 deletions .github/workflows/backend.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,6 +30,8 @@ jobs:
skip-drivers: ${{ matrix.skip-drivers }}
context: ${{ matrix.context }}
ubuntu-version: ${{ matrix.ubuntu-version }}
rocm-version: ${{ matrix.rocm-version || '6' }}
rocm-arch: ${{ matrix.rocm-arch || 'gfx1151' }}
secrets:
dockerUsername: ${{ secrets.DOCKERHUB_USERNAME }}
dockerPassword: ${{ secrets.DOCKERHUB_PASSWORD }}
Expand Down Expand Up @@ -1409,6 +1411,277 @@ jobs:
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
# ROCm 7.x hipblas builds (ubuntu:24.04 base + AMD apt repo; supports gfx1151/RDNA3.5+)
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-rerankers'
runs-on: 'ubuntu-latest'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "rerankers"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-llama-cpp'
runs-on: 'ubuntu-latest'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "llama-cpp"
dockerfile: "./backend/Dockerfile.llama-cpp"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-vllm'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "vllm"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-vllm-omni'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "vllm-omni"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-transformers'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "transformers"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-diffusers'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "diffusers"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-ace-step'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "ace-step"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-kokoro'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "kokoro"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-vibevoice'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "vibevoice"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-qwen-asr'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "qwen-asr"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-nemo'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "nemo"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-qwen-tts'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "qwen-tts"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-fish-speech'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "fish-speech"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-voxcpm'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "voxcpm"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-pocket-tts'
runs-on: 'arc-runner-set'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "pocket-tts"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-faster-whisper'
runs-on: 'bigger-runner'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "faster-whisper"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-whisperx'
runs-on: 'bigger-runner'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "whisperx"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
- build-type: 'hipblas'
cuda-major-version: ""
cuda-minor-version: ""
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-rocm7-hipblas-coqui'
runs-on: 'bigger-runner'
base-image: "ubuntu:24.04"
skip-drivers: 'false'
backend: "coqui"
dockerfile: "./backend/Dockerfile.python"
context: "./"
ubuntu-version: '2404'
rocm-version: '7'
rocm-arch: 'gfx1151'
# sycl builds
- build-type: 'intel'
cuda-major-version: ""
Expand Down
14 changes: 14 additions & 0 deletions .github/workflows/backend_build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -58,6 +58,16 @@ on:
required: false
default: '2204'
type: string
rocm-version:
description: 'ROCm major version (6 = ROCm 6.x with hipblas-dev/rocblas-dev, 7 = ROCm 7.x with amdrocm-* packages)'
required: false
default: '6'
type: string
rocm-arch:
description: 'ROCm GPU architecture for ROCm 7.x arch-specific packages (e.g. gfx1151)'
required: false
default: 'gfx1151'
type: string
secrets:
dockerUsername:
required: false
Expand Down Expand Up @@ -214,6 +224,8 @@ jobs:
BASE_IMAGE=${{ inputs.base-image }}
BACKEND=${{ inputs.backend }}
UBUNTU_VERSION=${{ inputs.ubuntu-version }}
ROCM_VERSION=${{ inputs.rocm-version }}
ROCM_ARCH=${{ inputs.rocm-arch }}
context: ${{ inputs.context }}
file: ${{ inputs.dockerfile }}
cache-from: type=gha
Expand All @@ -235,6 +247,8 @@ jobs:
BASE_IMAGE=${{ inputs.base-image }}
BACKEND=${{ inputs.backend }}
UBUNTU_VERSION=${{ inputs.ubuntu-version }}
ROCM_VERSION=${{ inputs.rocm-version }}
ROCM_ARCH=${{ inputs.rocm-arch }}
Copy link
Copy Markdown
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

is this specific to a single rocm arch? this is quite restrictive and would support only one specific version, which is not really what we want here. Is there a way to support a target list, like we do in llama.cpp?

AMDGPU_TARGETS?=gfx803,gfx900,gfx906,gfx908,gfx90a,gfx942,gfx1010,gfx1030,gfx1032,gfx1100,gfx1101,gfx1102,gfx1200,gfx1201

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Well it shouldn't be, but I ran into some trouble during the build so I was first narrowing it down to gfx1151 because it is one of the platforms actually benefitting from exactly the 7.12. rock. BUT that said, of course you are right and it should be for a list of AMDGPU_TARGETS. I will get on it.

context: ${{ inputs.context }}
file: ${{ inputs.dockerfile }}
cache-from: type=gha
Expand Down
15 changes: 15 additions & 0 deletions .github/workflows/image.yml
Original file line number Diff line number Diff line change
Expand Up @@ -29,6 +29,7 @@
makeflags: ${{ matrix.makeflags }}
ubuntu-version: ${{ matrix.ubuntu-version }}
ubuntu-codename: ${{ matrix.ubuntu-codename }}
rocm-version: ${{ matrix.rocm-version || '6' }}
secrets:
dockerUsername: ${{ secrets.DOCKERHUB_USERNAME }}
dockerPassword: ${{ secrets.DOCKERHUB_PASSWORD }}
Expand All @@ -47,6 +48,20 @@
makeflags: "--jobs=3 --output-sync=target"
ubuntu-version: '2404'
ubuntu-codename: 'noble'
rocm-version: '6'
# ROCm 7.x build for AMD Strix Halo / RDNA 3.5 (gfx1151) and other ROCm 7+ devices.
# Uses plain Ubuntu 24.04 and installs ROCm 7 from AMD's new apt repo (repo.amd.com).
- build-type: 'hipblas'
platforms: 'linux/amd64'
tag-latest: 'false'
tag-suffix: '-gpu-hipblas-rocm7'
base-image: "ubuntu:24.04"
grpc-base-image: "ubuntu:24.04"
runs-on: 'ubuntu-latest'
makeflags: "--jobs=3 --output-sync=target"
ubuntu-version: '2404'
ubuntu-codename: 'noble'
rocm-version: '7'

core-image-build:
if: github.repository == 'mudler/LocalAI'
Expand Down
7 changes: 7 additions & 0 deletions .github/workflows/image_build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -61,6 +61,11 @@ on:
required: false
default: 'noble'
type: string
rocm-version:
description: 'ROCm major version (6 or 7). Controls which apt packages are installed for hipblas builds.'
required: false
default: '6'
type: string
secrets:
dockerUsername:
required: true
Expand Down Expand Up @@ -217,6 +222,7 @@ jobs:
SKIP_DRIVERS=${{ inputs.skip-drivers }}
UBUNTU_VERSION=${{ inputs.ubuntu-version }}
UBUNTU_CODENAME=${{ inputs.ubuntu-codename }}
ROCM_VERSION=${{ inputs.rocm-version }}
context: .
file: ./Dockerfile
cache-from: type=gha
Expand Down Expand Up @@ -246,6 +252,7 @@ jobs:
SKIP_DRIVERS=${{ inputs.skip-drivers }}
UBUNTU_VERSION=${{ inputs.ubuntu-version }}
UBUNTU_CODENAME=${{ inputs.ubuntu-codename }}
ROCM_VERSION=${{ inputs.rocm-version }}
context: .
file: ./Dockerfile
cache-from: type=gha
Expand Down
Loading
Loading