llama.cpp/ggml
Zijun Yu b8690bc055 NPU Unify PD (#14)
* Stateless. Fix llama-cli llama-server

* Simplify broadcast op in attention

* Replace get_output_tensor+memcpy with set_output_tensor

* NPU unify PD. Unify dynamic and static dims
2026-01-15 11:27:30 -08:00
..
cmake ggml: Skip backend library linking code when GGML_BACKEND_DL=ON (#15094) 2025-08-07 13:45:41 +02:00
include Style: middle ptr and ref align, omit optional struct keyword 2026-01-15 11:27:30 -08:00
src NPU Unify PD (#14) 2026-01-15 11:27:30 -08:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt Refactor: clean, fix warning 2026-01-15 10:20:18 -08:00