HappyZ happyz
happyz synced commits to refs/pull/18892/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:40 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 24 commits »
happyz synced commits to refs/pull/18923/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:40 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 24 commits »
happyz synced commits to refs/pull/18858/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:39 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 8 commits »
happyz synced commits to refs/pull/18861/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:39 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 24 commits »
happyz synced commits to refs/pull/18872/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:39 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 34 commits »
happyz synced commits to refs/pull/18879/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:39 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 29 commits »
happyz synced commits to refs/pull/18886/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:39 -08:00
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
aa6f918c1c Vulkan Scalar Flash Attention Refactor (#19625)
Compare 7 commits »
happyz synced commits to refs/pull/18816/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:38 -08:00
244641955f models : fix graph splits (#19866)
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
Compare 5 commits »
happyz synced commits to refs/pull/18817/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:38 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 34 commits »
happyz synced commits to refs/pull/18836/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:38 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 27 commits »
happyz synced commits to refs/pull/18851/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:38 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 24 commits »
happyz synced commits to refs/pull/18711/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:37 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 6 commits »
happyz synced commits to refs/pull/18742/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:37 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 11 commits »
happyz synced commits to refs/pull/18745/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:37 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 23 commits »
happyz synced commits to refs/pull/18750/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:37 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 33 commits »
happyz synced commits to refs/pull/18576/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:36 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 11 commits »
happyz synced commits to refs/pull/18626/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:36 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 8 commits »
happyz synced commits to refs/pull/18633/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:36 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 29 commits »
happyz synced commits to refs/pull/18698/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:36 -08:00
47eb12b953 server: fix query params lost when proxying requests in multi-model router mode (#19854)
418dea39ce ggml/gguf : prevent integer overflows (#19856)
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
Compare 33 commits »
happyz synced commits to refs/pull/18059/merge at happyz/llama.cpp from mirror 2026-02-24 18:01:35 -08:00
fde17e0f5d webui: Client-side implementation of tool calling with calculator tool and (javascript) code interpreter tool
da426cb250 model : update label for LFM2-24B-A2B (#19848)
c830f99cfa server : support max_completion_tokens request property (#19831)
aa6f918c1c Vulkan Scalar Flash Attention Refactor (#19625)
Compare 10 commits »