..
peg-parser
common : introduce composable PEG parser combinators for chat parsing ( #17136 )
2025-12-03 12:45:32 +02:00
.gitignore
common : introduce composable PEG parser combinators for chat parsing ( #17136 )
2025-12-03 12:45:32 +02:00
CMakeLists.txt
common : introduce composable PEG parser combinators for chat parsing ( #17136 )
2025-12-03 12:45:32 +02:00
get-model.cpp
ci : add model tests + script wrapper ( #4586 )
2024-01-26 14:18:00 +02:00
get-model.h
ci : add model tests + script wrapper ( #4586 )
2024-01-26 14:18:00 +02:00
run-json-schema-to-grammar.mjs
llama : move end-user examples to tools directory ( #13249 )
2025-05-02 20:27:13 +02:00
test-alloc.cpp
ggml : fix graph reallocation with multiple chunks ( #16396 )
2025-10-03 13:49:08 +02:00
test-arg-parser.cpp
arg: fix common_params_parse not accepting negated arg ( #17991 )
2025-12-13 12:53:37 +01:00
test-autorelease.cpp
llama : add `llama_vocab`, functions -> methods, naming ( #11110 )
2025-01-12 11:32:42 +02:00
test-backend-ops.cpp
vulkan: Multi-pass softmax for large number of cols ( #17892 )
2025-12-13 10:04:29 +01:00
test-barrier.cpp
Fix race conditions in threadpool when dealing with dynamic/frequent n_threads changes ( #17748 )
2025-12-10 12:32:23 -08:00
test-c.c
ggml : remove kompute backend ( #14501 )
2025-07-03 07:48:32 +03:00
test-chat-parser.cpp
common : handle unicode during partial json parsing ( #16526 )
2025-10-12 16:18:47 +03:00
test-chat-peg-parser.cpp
common : introduce composable PEG parser combinators for chat parsing ( #17136 )
2025-12-03 12:45:32 +02:00
test-chat-template.cpp
chat : Granite Docling stopping ( #16438 )
2025-10-06 18:59:40 +02:00
test-chat.cpp
chat-parser: handle whitespace around JSON in tool call parsing
2025-12-14 23:54:41 +00:00
test-double-float.cpp
ggml : minor naming changes ( #8433 )
2024-07-12 10:46:02 +03:00
test-gbnf-validator.cpp
cmake : do not include ./src as public for libllama ( #13062 )
2025-04-24 16:00:10 +03:00
test-gguf.cpp
gguf: fix failure on version == 0 ( #13956 )
2025-06-01 18:08:05 +02:00
test-grammar-integration.cpp
llama : add token matching support to llama-grammar ( #17816 )
2025-12-09 00:32:57 -06:00
test-grammar-llguidance.cpp
cmake : do not include ./src as public for libllama ( #13062 )
2025-04-24 16:00:10 +03:00
test-grammar-parser.cpp
llama : add token matching support to llama-grammar ( #17816 )
2025-12-09 00:32:57 -06:00
test-json-partial.cpp
common : handle unicode during partial json parsing ( #16526 )
2025-10-12 16:18:47 +03:00
test-json-schema-to-grammar.cpp
Server: Change Invalid Schema from Server Error (500) to User Error (400) ( #17572 )
2025-12-02 17:33:50 +01:00
test-llama-grammar.cpp
llama : add token matching support to llama-grammar ( #17816 )
2025-12-09 00:32:57 -06:00
test-log.cpp
common : use common_ prefix for common library functions ( #9805 )
2024-10-10 22:57:42 +02:00
test-lora-conversion-inference.sh
cli: new CLI experience ( #17824 )
2025-12-10 15:28:59 +01:00
test-model-load-cancel.cpp
llama : update llama_model API names ( #11063 )
2025-01-06 10:55:18 +02:00
test-mtmd-c-api.c
mtmd : add C public API ( #13184 )
2025-05-04 23:43:42 +02:00
test-opt.cpp
tests : fix test-opt with GGML_BACKEND_DL ( #15599 )
2025-08-26 22:14:38 +02:00
test-peg-parser.cpp
common : introduce composable PEG parser combinators for chat parsing ( #17136 )
2025-12-03 12:45:32 +02:00
test-quantize-fns.cpp
tests : fix test-quantize-fns to init the CPU backend ( #12306 )
2025-03-10 14:07:15 +02:00
test-quantize-perf.cpp
ci: run the x64 and arm ci on the github machines instead ( #16183 )
2025-09-25 08:06:06 +03:00
test-quantize-stats.cpp
server: introduce API for serving / loading / unloading multiple models ( #17470 )
2025-12-01 19:41:04 +01:00
test-regex-partial.cpp
`common`: add partial regex support ( #12808 )
2025-05-14 19:50:57 +01:00
test-rope.cpp
ggml-cpu: templateify ggml_compute_forward_rope_f32 and _f16 ( #16805 )
2025-11-11 13:33:24 +02:00
test-sampling.cpp
sampling : optimize samplers by reusing bucket sort ( #15665 )
2025-08-31 20:41:02 +03:00
test-thread-safety.cpp
server : support unified cache across slots ( #16736 )
2025-11-02 18:14:04 +02:00
test-tokenizer-0.cpp
llama : add `llama_vocab`, functions -> methods, naming ( #11110 )
2025-01-12 11:32:42 +02:00
test-tokenizer-0.py
py : logging and flake8 suppression refactoring ( #7081 )
2024-05-05 08:07:48 +03:00
test-tokenizer-0.sh
scripts : make the shell scripts cross-platform ( #14341 )
2025-06-30 10:17:18 +02:00
test-tokenizer-1-bpe.cpp
cmake : do not include ./src as public for libllama ( #13062 )
2025-04-24 16:00:10 +03:00
test-tokenizer-1-spm.cpp
cmake : do not include ./src as public for libllama ( #13062 )
2025-04-24 16:00:10 +03:00
test-tokenizer-random.py
requirements : update transformers/torch for Embedding Gemma ( #15828 )
2025-09-09 06:06:52 +02:00
test-tokenizers-repo.sh
devops: add s390x & ppc64le CI ( #15925 )
2025-09-27 02:03:33 +08:00