diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 8161d08d53..fc19bb9c84 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -82,6 +82,7 @@ jobs: sysctl -a cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DCMAKE_BUILD_RPATH="@loader_path" \ -DLLAMA_FATAL_WARNINGS=ON \ -DLLAMA_BUILD_BORINGSSL=ON \ @@ -121,6 +122,7 @@ jobs: # https://github.com/ggml-org/llama.cpp/actions/runs/8635935781/job/23674807267#step:5:2313 cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DCMAKE_BUILD_RPATH="@loader_path" \ -DLLAMA_FATAL_WARNINGS=ON \ -DLLAMA_BUILD_BORINGSSL=ON \ @@ -168,7 +170,7 @@ jobs: id: cmake_build run: | export CMAKE_PREFIX_PATH=dawn - cmake -B build -G "Ninja" -DGGML_WEBGPU=ON -DGGML_METAL=OFF -DGGML_BLAS=OFF + cmake -B build -G "Ninja" -DCMAKE_BUILD_TYPE=Release -DGGML_WEBGPU=ON -DGGML_METAL=OFF -DGGML_BLAS=OFF cmake --build build --config Release -j $(sysctl -n hw.logicalcpu) - name: Test @@ -233,6 +235,7 @@ jobs: run: | cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DLLAMA_FATAL_WARNINGS=ON \ -DGGML_RPC=ON cmake --build build --config Release -j $(nproc) @@ -285,6 +288,7 @@ jobs: run: | cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DGGML_RPC=ON cmake --build build --config Release -j $(nproc) @@ -385,6 +389,7 @@ jobs: export Dawn_DIR=dawn/lib64/cmake/Dawn cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DGGML_WEBGPU=ON cmake --build build --config Release -j $(nproc) @@ -424,6 +429,7 @@ jobs: source emsdk/emsdk_env.sh emcmake cmake -B build-wasm \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DGGML_WEBGPU=ON \ -DLLAMA_OPENSSL=OFF \ -DEMDAWNWEBGPU_DIR=emdawnwebgpu_pkg @@ -457,6 +463,7 @@ jobs: run: | cmake -B build -S . \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DCMAKE_HIP_COMPILER="$(hipconfig -l)/clang" \ -DGGML_HIP_ROCWMMA_FATTN=ON \ -DGGML_HIP=ON @@ -489,6 +496,7 @@ jobs: run: | cmake -B build -S . \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DGGML_MUSA=ON cmake --build build --config Release -j $(nproc) @@ -537,6 +545,7 @@ jobs: source /opt/intel/oneapi/setvars.sh cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DGGML_SYCL=ON \ -DCMAKE_C_COMPILER=icx \ -DCMAKE_CXX_COMPILER=icpx @@ -587,6 +596,7 @@ jobs: source /opt/intel/oneapi/setvars.sh cmake -B build \ -G "Ninja" \ + -DCMAKE_BUILD_TYPE=Release \ -DGGML_SYCL=ON \ -DCMAKE_C_COMPILER=icx \ -DCMAKE_CXX_COMPILER=icpx \ @@ -692,7 +702,7 @@ jobs: defines: '-G "Ninja Multi-Config" -D CMAKE_TOOLCHAIN_FILE=cmake/x64-windows-llvm.cmake -DGGML_NATIVE=OFF -DLLAMA_BUILD_SERVER=ON -DGGML_RPC=ON -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON -DGGML_OPENMP=OFF -DGGML_BLAS=ON -DGGML_BLAS_VENDOR=OpenBLAS -DBLAS_INCLUDE_DIRS="$env:RUNNER_TEMP/openblas/include" -DBLAS_LIBRARIES="$env:RUNNER_TEMP/openblas/lib/openblas.lib"' - build: 'vulkan-x64' arch: 'x64' - defines: '-G "Ninja Multi-Config" -DCMAKE_BUILD_TYPE=Release -DGGML_NATIVE=OFF -DLLAMA_BUILD_SERVER=ON -DGGML_RPC=ON -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON -DGGML_VULKAN=ON' + defines: '-G "Ninja" -DCMAKE_BUILD_TYPE=Release -DGGML_NATIVE=OFF -DLLAMA_BUILD_SERVER=ON -DGGML_RPC=ON -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON -DGGML_VULKAN=ON' - build: 'llvm-arm64' arch: 'arm64' defines: '-G "Ninja Multi-Config" -D CMAKE_TOOLCHAIN_FILE=cmake/arm64-windows-llvm.cmake -DGGML_NATIVE=OFF -DLLAMA_BUILD_SERVER=ON' @@ -867,7 +877,7 @@ jobs: # TODO: Remove GGML_CUDA_CUB_3DOT2 flag once CCCL 3.2 is bundled within CTK and that CTK version is used in this project run: | call "C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Auxiliary\Build\vcvarsall.bat" x64 - cmake -S . -B build -G "Ninja" ^ + cmake -S . -B build -G "Ninja Multi-Config" ^ -DLLAMA_BUILD_SERVER=ON ^ -DLLAMA_BUILD_BORINGSSL=ON ^ -DGGML_NATIVE=OFF ^