convert : DSA indexer weights are bf16 in the original fp8 model, so I think it's best not to quantize them.
This commit is contained in:
parent
72b7214467
commit
961bc95d96
|
|
@ -618,6 +618,8 @@ class ModelBase:
|
|||
gguf.MODEL_TENSOR.SSM_CONV1D_Q,
|
||||
gguf.MODEL_TENSOR.SSM_CONV1D_K,
|
||||
gguf.MODEL_TENSOR.SSM_CONV1D_V,
|
||||
# DSA indexer weights should be F32
|
||||
gguf.MODEL_TENSOR.INDEXER_PROJ,
|
||||
)
|
||||
)
|
||||
or new_name[-7:] not in (".weight", ".lora_a", ".lora_b")
|
||||
|
|
|
|||
Loading…
Reference in New Issue