From 48cda24c1175363dd17925102dbe1da49279940e Mon Sep 17 00:00:00 2001 From: AN Long Date: Fri, 27 Mar 2026 19:36:13 +0800 Subject: [PATCH] server: remove the verbose_prompt parameter (#21059) * server: respect the verbose_prompt parameter * Revert "server: respect the verbose_prompt parameter" This reverts commit 8ed885cf375b2c8ba641c661f3667df70b9797f4. * Remove --verbose-prompt parameter from llama-server * Using set_examples instead of set_excludes --- common/arg.cpp | 2 +- tools/server/README.md | 1 - 2 files changed, 1 insertion(+), 2 deletions(-) diff --git a/common/arg.cpp b/common/arg.cpp index 32d8e69f96..5bab9abc77 100644 --- a/common/arg.cpp +++ b/common/arg.cpp @@ -1079,7 +1079,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex [](common_params & params) { params.verbose_prompt = true; } - )); + ).set_examples({LLAMA_EXAMPLE_COMPLETION, LLAMA_EXAMPLE_CLI, LLAMA_EXAMPLE_EMBEDDING, LLAMA_EXAMPLE_RETRIEVAL})); add_opt(common_arg( {"--display-prompt"}, {"--no-display-prompt"}, diff --git a/tools/server/README.md b/tools/server/README.md index cb53678416..f99103a584 100644 --- a/tools/server/README.md +++ b/tools/server/README.md @@ -36,7 +36,6 @@ For the full list of features, please refer to [server's changelog](https://gith | `--license` | show source code license and dependencies | | `-cl, --cache-list` | show list of models in cache | | `--completion-bash` | print source-able bash completion script for llama.cpp | -| `--verbose-prompt` | print a verbose prompt before generation (default: false) | | `-t, --threads N` | number of CPU threads to use during generation (default: -1)
(env: LLAMA_ARG_THREADS) | | `-tb, --threads-batch N` | number of threads to use during batch and prompt processing (default: same as --threads) | | `-C, --cpu-mask M` | CPU affinity mask: arbitrarily long hex. Complements cpu-range (default: "") |