llama.cpp/ggml
Oliver Simons 559d058dd2 CUDA: Move cccl fetch to after cuda has been enabled in CMakeLists.txt
This will allow cccl to set build flags for the CUDA compiler, required
e.g. for MSVC compat, see also
https://github.com/NVIDIA/cccl/pull/6791
2025-12-02 11:23:14 +01:00
..
cmake ggml: Skip backend library linking code when GGML_BACKEND_DL=ON (#15094) 2025-08-07 13:45:41 +02:00
include model: LFM2-VL fixes (#17577) 2025-11-30 21:57:31 +01:00
src CUDA: Move cccl fetch to after cuda has been enabled in CMakeLists.txt 2025-12-02 11:23:14 +01:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt ggml : add GGML_SCHED_NO_REALLOC option to disable reallocations in ggml_backend_sched (#17276) 2025-11-28 17:33:23 +02:00