llama.cpp/gguf-py/gguf
Sigbjørn Skjæret 06808a3d0d
Support converting models with multiple chat templates
Adds the following metadata:
* tokenizer.chat_templates
* tokenizer.chat_template.<name1>
* tokenizer.chat_template.<name2>
* tokenizer.chat_template.<...>

Where `tokenizer.chat_templates` is an array of the template names (except `default`), `default` is added to the regular `tokenizer.chat_template`.
2024-04-10 15:31:05 +02:00
..
__init__.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
constants.py Support converting models with multiple chat templates 2024-04-10 15:31:05 +02:00
gguf.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
gguf_reader.py gguf : add support for I64 and F64 arrays (#6062) 2024-03-15 10:46:51 +02:00
gguf_writer.py Support converting models with multiple chat templates 2024-04-10 15:31:05 +02:00
py.typed convert : various script cleanups/fixes + merges and special token handling (#2842) 2023-08-30 11:25:50 +03:00
tensor_mapping.py llama : add Command R Plus support (#6491) 2024-04-09 11:16:13 +03:00
vocab.py Support converting models with multiple chat templates 2024-04-10 15:31:05 +02:00