llama.cpp/docs/preset.md

1.5 KiB

llama.cpp INI Presets

Introduction

The INI preset feature, introduced in PR#17859, allows users to create reusable and shareable parameter configurations for llama.cpp.

Using Presets with the Server

When running multiple models on the server (router mode), INI preset files can be used to configure model-specific parameters. Please refer to the server documentation for more details.

Using a Remote Preset

[!NOTE]

This feature is currently only supported via the -hf option.

For GGUF models hosted on Hugging Face, you can include a preset.ini file in the root directory of the repository to define specific configurations for that model.

Example:

hf-repo-draft = username/my-draft-model-GGUF
temp = 0.5
top-k = 20
top-p = 0.95

For security reasons, only certain options are allowed. Please refer to preset.cpp for the complete list of permitted options.

Example usage:

Assuming your repository username/my-model-with-preset contains a preset.ini with the configuration above:

llama-cli -hf username/my-model-with-preset

# This is equivalent to:
llama-cli -hf username/my-model-with-preset \
  --hf-repo-draft username/my-draft-model-GGUF \
  --temp 0.5 \
  --top-k 20 \
  --top-p 0.95

You can also override preset arguments by specifying them on the command line:

# Force temp = 0.1, overriding the preset value
llama-cli -hf username/my-model-with-preset --temp 0.1