llama.cpp/tools
Georgi Gerganov 7b50d589a8
kv-cells : fix tracking of seq_pos (#14339)
* kv-cells : fix tracking of seq_pos during cache reuse

ggml-ci

* cont : improve error message

ggml-ci

* cont : add more comments
2025-06-23 12:27:35 +03:00
..
batched-bench
cvector-generator
export-lora
gguf-split
imatrix
llama-bench llama-bench : add --no-warmup flag (#14224) (#14270) 2025-06-19 12:24:12 +02:00
main
mtmd mtmd : fix Pixtral OOM with large images by capping image_size to 1024 (#14326) 2025-06-22 14:44:57 +02:00
perplexity
quantize quantize : handle user-defined pruning of whole layers (blocks) (#13037) 2025-06-22 23:16:26 +02:00
rpc
run run : avoid double tokenization (#14327) 2025-06-23 01:28:06 +08:00
server kv-cells : fix tracking of seq_pos (#14339) 2025-06-23 12:27:35 +03:00
tokenize
tts
CMakeLists.txt