From 68a1c4dc51196792ab49fa0117013b27b2f4c5d7 Mon Sep 17 00:00:00 2001 From: Daniel Bevenius Date: Wed, 17 Dec 2025 09:00:46 +0100 Subject: [PATCH] llama : clarify backend_accept/backend_set_input comments [no ci] --- include/llama.h | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/include/llama.h b/include/llama.h index 5ba9d6cb32..034bb4e39b 100644 --- a/include/llama.h +++ b/include/llama.h @@ -1215,7 +1215,7 @@ extern "C" { // note: call once per sampler bool (*backend_init)(struct llama_sampler * smpl, ggml_backend_buffer_type_t buft); - // call after .backend_accept() + // call after .backend_apply() void (*backend_accept)( struct llama_sampler * smpl, struct ggml_context * ctx, @@ -1229,7 +1229,7 @@ extern "C" { struct ggml_cgraph * gf, struct llama_sampler_data * data); - // call before .backend_apply() + // called before graph execution to set inputs for the current ubatch void (*backend_set_input)(struct llama_sampler * smpl); };