llama.cpp/ggml
Gaurav Garg aa8b62105c Support device-specific host buffer types if all underlying backends expose the same type. This allows using pinned memory instead of pageable memory for CUDA.
Fix compilation errors.
2026-02-16 15:39:26 +05:30
..
cmake Remove shfl and AllReduce from backend interface 2026-02-11 14:51:37 +01:00
include support for tensor dims % n_devs != 0 2026-02-13 00:40:00 +01:00
src Support device-specific host buffer types if all underlying backends expose the same type. This allows using pinned memory instead of pageable memory for CUDA. 2026-02-16 15:39:26 +05:30
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt GGML: HIP: add RCCL support 2026-02-11 14:51:33 +01:00