llama.cpp/include
Quentin Fuxa 42c29c3970 llama : add attention weights extraction API [EXPERIMENTAL] 2026-03-31 22:13:17 +02:00
..
llama-cpp.h llama : re-enable manual LoRA adapter free (#19983) 2026-03-18 12:03:26 +02:00
llama.h llama : add attention weights extraction API [EXPERIMENTAL] 2026-03-31 22:13:17 +02:00