fix: llama arch implementation (#17665)

This commit is contained in:
Gilad S. 2025-12-01 22:21:13 +02:00 committed by GitHub
parent ec18edfcba
commit 00c361fe53
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
1 changed files with 2 additions and 0 deletions

View File

@ -626,6 +626,8 @@ void llama_model::load_hparams(llama_model_loader & ml) {
switch (arch) { switch (arch) {
case LLM_ARCH_LLAMA: case LLM_ARCH_LLAMA:
{ {
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
if (hparams.n_expert == 8) { if (hparams.n_expert == 8) {
switch (hparams.n_layer) { switch (hparams.n_layer) {
case 32: type = LLM_TYPE_8x7B; break; case 32: type = LLM_TYPE_8x7B; break;