diff --git a/src/llama-model.cpp b/src/llama-model.cpp index 84ac4d3a9e..50900feb2c 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -6853,7 +6853,6 @@ bool llama_model::load_tensors(llama_model_loader & ml) { // o_norm (reusing SSM_NORM) layer.ssm_o_norm = create_tensor(tn(LLM_TENSOR_SSM_NORM, "weight", i), {n_embd_head_k_kda}, 0); // FusedRMSNormGated - layer.ssm_o_norm_b = create_tensor(tn(LLM_TENSOR_SSM_NORM, "bias", i), {n_embd_head_k_kda}, TENSOR_NOT_REQUIRED); // o_proj layer.wo = create_tensor(tn(LLM_TENSOR_ATTN_OUT, "weight", i), {n_embd_head_v_kda * n_head, n_embd}, 0); diff --git a/src/llama-model.h b/src/llama-model.h index 359701589c..5b408bcea2 100644 --- a/src/llama-model.h +++ b/src/llama-model.h @@ -423,7 +423,6 @@ struct llama_layer { struct ggml_tensor * ssm_g_a = nullptr; struct ggml_tensor * ssm_g_b = nullptr; struct ggml_tensor * ssm_o_norm = nullptr; - struct ggml_tensor * ssm_o_norm_b = nullptr; struct llama_layer_posnet posnet;