From aba181ebadd1c860201eff4fe702a89b3c9b8a1c Mon Sep 17 00:00:00 2001 From: Yee Man Chan Date: Mon, 5 Jan 2026 19:21:06 +0800 Subject: [PATCH] removed LOG_INFO --- src/models/kimi-linear.cpp | 3 --- 1 file changed, 3 deletions(-) diff --git a/src/models/kimi-linear.cpp b/src/models/kimi-linear.cpp index 32a723b80a..a943dd1dce 100644 --- a/src/models/kimi-linear.cpp +++ b/src/models/kimi-linear.cpp @@ -1,6 +1,5 @@ #include "models.h" #include "ggml.h" -#include "llama-impl.h" #define CHUNK_SIZE 64 @@ -58,8 +57,6 @@ llm_build_kimi_linear::llm_build_kimi_linear(const llama_model & model, const ll const auto & layer = model.layers[il]; ggml_tensor * inpSA = inpL; - if (!layer.attn_norm) - LLAMA_LOG_INFO("Empty attn_norm at layer %d\n", il); // Attention Norm cur = build_norm(inpL, layer.attn_norm, NULL, LLM_NORM_RMS, il); cb(cur, "attn_norm", il);