llama.cpp/gguf-py/gguf
Sigbjørn Skjæret b8e09f08b9
model : add grok-2 support (#15539)
* add grok-2 support

* type fix

* type fix

* type fix

* "fix" vocab for invalid sequences

* fix expert tensor mapping and spaces in vocab

* add chat template

* fix norm tensor mapping

* rename layer_out_norm to ffn_post_norm

* ensure ffn_post_norm is mapped

* fix experts merging

* remove erroneous FFN_GATE entry

* concatenate split tensors and add more metadata

* process all expert layers and try cat instead of hstack

* add support for community BPE vocab

* fix expert feed forward length and ffn_down concat

* commit this too

* add ffn_up/gate/down, unsure if sequence is right

* add ffn_gate/down/up to tensor names

* correct residual moe (still not working)

* mess--

* fix embedding scale being applied twice

* add built in chat template

* change beta fast for grok if default value

* remove spm vocab in favor of community bpe vocab

* change attention temp length metadata type to integer

* update attention temp length metadata

* remove comment

* replace M_SQRT2 with std::sqrt(2)

* add yarn metadata, move defaults to hparams
2025-09-14 23:00:59 +02:00
..
scripts gguf-py: byteswapping improvements (#12851) 2025-08-28 16:56:41 +08:00
__init__.py convert-*.py: GGUF Naming Convention Refactor and Metadata Override Refactor (#7499) 2024-07-18 20:40:15 +10:00
constants.py model : add grok-2 support (#15539) 2025-09-14 23:00:59 +02:00
gguf.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
gguf_reader.py gguf-py : display the invalid gguf type (#13687) 2025-05-21 16:33:54 +02:00
gguf_writer.py model : add grok-2 support (#15539) 2025-09-14 23:00:59 +02:00
lazy.py gguf-py : support lazy tensor splitting (#12809) 2025-04-08 09:03:07 +02:00
metadata.py ggml : model card yaml tab->2xspace (#14819) 2025-07-22 19:29:43 +03:00
py.typed convert : various script cleanups/fixes + merges and special token handling (#2842) 2023-08-30 11:25:50 +03:00
quants.py gguf-py : add Numpy MXFP4 de/quantization support (#15111) 2025-08-08 17:48:26 -04:00
tensor_mapping.py model : add grok-2 support (#15539) 2025-09-14 23:00:59 +02:00
utility.py convert : improve Mistral models integration (#14737) 2025-08-11 10:07:49 +02:00
vocab.py model : add text-only support for Kimi-VL (and find special tokens in text_config) (#15051) 2025-08-03 16:56:25 +02:00