From 8a361194075fed700e2ac0d813c44213a454036e Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Fri, 2 Jan 2026 16:00:25 +0200 Subject: [PATCH] cont : fix --- src/llama-context.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama-context.cpp b/src/llama-context.cpp index ffdb487402..939ca98948 100644 --- a/src/llama-context.cpp +++ b/src/llama-context.cpp @@ -342,7 +342,7 @@ llama_context::llama_context( reserve(); - if (cparams.flash_attn) { + if (!cparams.flash_attn) { if (ggml_is_quantized(params.type_v)) { throw std::runtime_error("quantized V cache was requested, but this requires Flash Attention"); }