convert : DSA indexer weights are bf16 in the original fp8 model, so I think it's best not to quantize them.

This commit is contained in:
Stanisław Szymczyk 2026-03-14 20:17:21 +01:00
parent 72b7214467
commit 961bc95d96
1 changed files with 2 additions and 0 deletions

View File

@ -618,6 +618,8 @@ class ModelBase:
gguf.MODEL_TENSOR.SSM_CONV1D_Q,
gguf.MODEL_TENSOR.SSM_CONV1D_K,
gguf.MODEL_TENSOR.SSM_CONV1D_V,
# DSA indexer weights should be F32
gguf.MODEL_TENSOR.INDEXER_PROJ,
)
)
or new_name[-7:] not in (".weight", ".lora_a", ".lora_b")