fixed vram leak

This commit is contained in:
Aaron Lee 2025-08-19 23:10:56 -04:00
parent d72f9d5691
commit 471e026327
1 changed files with 2 additions and 0 deletions

View File

@ -3050,5 +3050,7 @@ void llama_build_and_execute_mtp_graph(struct llama_context * ctx,
if (logits_mtp) {
ctx->set_logits_ith(logits_mtp, sched, last_tok_idx);
}
ggml_backend_sched_free(sched);
}