llama.cpp/tools
Francis Couture-Harpin 50f53b3e40 imatrix : warn when writing partial data, to help guess dataset coverage
Also make the legacy format store partial data
by using neutral values for missing data.
This matches what is done at read-time for the new format,
and so should get the same quality in case the old format is still used.
2025-07-12 16:50:10 -04:00
..
batched-bench llama : deprecate llama_kv_self_ API (#14030) 2025-06-06 14:11:15 +03:00
cvector-generator llama : deprecate llama_kv_self_ API (#14030) 2025-06-06 14:11:15 +03:00
export-lora
gguf-split scripts : make the shell scripts cross-platform (#14341) 2025-06-30 10:17:18 +02:00
imatrix imatrix : warn when writing partial data, to help guess dataset coverage 2025-07-12 16:50:10 -04:00
llama-bench llama-bench : add --no-warmup flag (#14224) (#14270) 2025-06-19 12:24:12 +02:00
main main : honor --verbose-prompt on interactive prompts (#14350) 2025-06-24 09:31:00 +02:00
mtmd ggml : implement GEGLU_ERF and GEGLU_QUICK ops (#14445) 2025-07-03 23:07:22 +02:00
perplexity llama : deprecate llama_kv_self_ API (#14030) 2025-06-06 14:11:15 +03:00
quantize Merge branch 'master' into compilade/imatrix-batched-chunks 2025-07-12 13:31:19 -04:00
rpc rpc : Fix build on OpenBSD (#13541) 2025-05-25 15:35:53 +03:00
run cmake : do not search for curl libraries by ourselves (#14613) 2025-07-10 15:29:05 +03:00
server server : fix pooled embedding output (#14645) 2025-07-12 13:21:02 +03:00
tokenize
tts sync : vendor (#13901) 2025-05-30 16:25:45 +03:00
CMakeLists.txt mtmd : rename llava directory to mtmd (#13311) 2025-05-05 16:02:55 +02:00