llama.cpp/gguf-py/gguf
Xuan-Son Nguyen dbc15a7967
convert: support Mistral 3 Large MoE (#17730)
* convert: support Mistral 3 Large MoE

* filter out vision tensors, add missing keys

* handle vocab

* add temperature_length

* fix mscale_all_dim

* clean up

* Apply suggestions from code review

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>

* fix

* Update gguf-py/gguf/tensor_mapping.py

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>

---------

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>
2025-12-06 10:49:33 +01:00
..
scripts gguf-py : fix passing non-native endian tensors (editor-gui and new-metadata) (#17553) 2025-11-28 20:53:01 +01:00
__init__.py
constants.py model: support Ministral3 (#17644) 2025-12-01 12:26:52 +01:00
gguf.py
gguf_reader.py
gguf_writer.py model: support Ministral3 (#17644) 2025-12-01 12:26:52 +01:00
lazy.py convert : handle compressed-tensors quant method (#17069) 2025-11-09 09:45:50 -05:00
metadata.py llama: introduce support for model-embedded sampling parameters (#17120) 2025-11-25 09:56:07 +08:00
py.typed
quants.py gguf-py : add Numpy MXFP4 de/quantization support (#15111) 2025-08-08 17:48:26 -04:00
tensor_mapping.py convert: support Mistral 3 Large MoE (#17730) 2025-12-06 10:49:33 +01:00
utility.py convert : parse safetensors directly (#15667) 2025-11-09 09:49:40 -05:00
vocab.py convert : support latest mistral-common (fix conversion with --mistral-format) (#17712) 2025-12-03 21:15:04 +01:00