From 9fe9a00a8a778f60df2d4ea6e435863b24694695 Mon Sep 17 00:00:00 2001 From: Daniel Bevenius Date: Mon, 17 Nov 2025 15:30:16 +0100 Subject: [PATCH] llama-cli : add backend sampler configuration --- tools/main/main.cpp | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tools/main/main.cpp b/tools/main/main.cpp index 33e8862335..f1d0fd4b60 100644 --- a/tools/main/main.cpp +++ b/tools/main/main.cpp @@ -147,6 +147,9 @@ int main(int argc, char ** argv) { return 1; } + // Configure backend sampler chain + llama_set_backend_sampler(ctx, 0, common_sampler_backend_init(model, sparams)); + auto * mem = llama_get_memory(ctx); const llama_vocab * vocab = llama_model_get_vocab(model);