From 7dba049b0707ae395c59b085c5fd52cae7b74fe0 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sigbj=C3=B8rn=20Skj=C3=A6ret?= Date: Thu, 4 Dec 2025 11:25:08 +0100 Subject: [PATCH] ci : disable ggml-ci-x64-amd-* (#17753) --- .github/workflows/build.yml | 44 ++++++++++++++++++------------------- 1 file changed, 22 insertions(+), 22 deletions(-) diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 49e836d9b2..ad205f3ec9 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -1602,33 +1602,33 @@ jobs: run: | bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp - ggml-ci-x64-amd-vulkan: - runs-on: [self-hosted, Linux, X64, AMD] + # ggml-ci-x64-amd-vulkan: + # runs-on: [self-hosted, Linux, X64, AMD] - steps: - - name: Clone - id: checkout - uses: actions/checkout@v4 + # steps: + # - name: Clone + # id: checkout + # uses: actions/checkout@v4 - - name: Test - id: ggml-ci - run: | - vulkaninfo --summary - GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp + # - name: Test + # id: ggml-ci + # run: | + # vulkaninfo --summary + # GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp - ggml-ci-x64-amd-rocm: - runs-on: [self-hosted, Linux, X64, AMD] + # ggml-ci-x64-amd-rocm: + # runs-on: [self-hosted, Linux, X64, AMD] - steps: - - name: Clone - id: checkout - uses: actions/checkout@v4 + # steps: + # - name: Clone + # id: checkout + # uses: actions/checkout@v4 - - name: Test - id: ggml-ci - run: | - amd-smi static - GG_BUILD_ROCM=1 GG_BUILD_AMDGPU_TARGETS="gfx1101" bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp + # - name: Test + # id: ggml-ci + # run: | + # amd-smi static + # GG_BUILD_ROCM=1 GG_BUILD_AMDGPU_TARGETS="gfx1101" bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp ggml-ci-mac-metal: runs-on: [self-hosted, macOS, ARM64]