refactor: remove rope_parameters from VaetkiModel
This commit is contained in:
parent
75323b3e08
commit
bfc92e954b
|
|
@ -7733,11 +7733,6 @@ class VaetkiModel(TextModel):
|
||||||
self.gguf_writer.add_value_length_mla(hparams["v_head_dim"])
|
self.gguf_writer.add_value_length_mla(hparams["v_head_dim"])
|
||||||
self.gguf_writer.add_rope_dimension_count(hparams["qk_rope_head_dim"])
|
self.gguf_writer.add_rope_dimension_count(hparams["qk_rope_head_dim"])
|
||||||
|
|
||||||
self.rope_parameters = {
|
|
||||||
"full_attention": {"rope_theta": self.hparams.get("rope_theta_global", 1000000.0)},
|
|
||||||
"sliding_attention": {"rope_theta": self.hparams.get("rope_theta", 10000.0)}
|
|
||||||
}
|
|
||||||
|
|
||||||
# MoE parameters
|
# MoE parameters
|
||||||
self.gguf_writer.add_leading_dense_block_count(hparams.get("first_k_dense_replace", 1))
|
self.gguf_writer.add_leading_dense_block_count(hparams.get("first_k_dense_replace", 1))
|
||||||
self.gguf_writer.add_expert_count(hparams["n_routed_experts"])
|
self.gguf_writer.add_expert_count(hparams["n_routed_experts"])
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue