llama : fix build_ffn without gate
This commit is contained in:
parent
15a28ec8c7
commit
98ce93e776
|
|
@ -782,7 +782,7 @@ ggml_tensor * llm_graph_context::build_ffn(
|
||||||
} break;
|
} break;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (type_gate == LLM_FFN_PAR) {
|
if (gate && type_gate == LLM_FFN_PAR) {
|
||||||
cur = ggml_mul(ctx0, cur, tmp);
|
cur = ggml_mul(ctx0, cur, tmp);
|
||||||
cb(cur, "ffn_gate_par", il);
|
cb(cur, "ffn_gate_par", il);
|
||||||
}
|
}
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue