llama.cpp/gguf-py/gguf
Aman Gupta 8a4a856277
Add LLaDA 8b Diffusion model (#14771)
* Add support for Llada-8b: diffusion model

* Add README

* Fix README and convert_hf_to_gguf

* convert_hf_to_gguf.py: address review comments

* Make everything in a single example

* Remove model-specific sampling

* Remove unused argmax

* Remove braced initializers, improve README.md a bit

* Add diffusion specific gguf params in set_vocab, remove setting rope_theta and rms_norm_eps

* Remove adding the mask token

* Move add_add_bos_token to set_vocab

* use add_bool in gguf_writer.py
2025-07-31 19:49:09 +08:00
..
scripts gguf-py : dump bpw per layer and model in markdown mode (#14703) 2025-07-16 00:04:42 +02:00
__init__.py convert-*.py: GGUF Naming Convention Refactor and Metadata Override Refactor (#7499) 2024-07-18 20:40:15 +10:00
constants.py Add LLaDA 8b Diffusion model (#14771) 2025-07-31 19:49:09 +08:00
gguf.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
gguf_reader.py gguf-py : display the invalid gguf type (#13687) 2025-05-21 16:33:54 +02:00
gguf_writer.py Add LLaDA 8b Diffusion model (#14771) 2025-07-31 19:49:09 +08:00
lazy.py gguf-py : support lazy tensor splitting (#12809) 2025-04-08 09:03:07 +02:00
metadata.py ggml : model card yaml tab->2xspace (#14819) 2025-07-22 19:29:43 +03:00
py.typed convert : various script cleanups/fixes + merges and special token handling (#2842) 2023-08-30 11:25:50 +03:00
quants.py ggml-quants : ternary packing for TriLMs and BitNet b1.58 (#8151) 2024-09-05 21:48:47 -04:00
tensor_mapping.py Add LLaDA 8b Diffusion model (#14771) 2025-07-31 19:49:09 +08:00
utility.py gguf-py : fix SafetensorRemote return on undefined size (< 0) (#13841) 2025-05-28 23:50:20 +02:00
vocab.py mtmd : add support for Voxtral (#14862) 2025-07-28 15:01:48 +02:00