llama.cpp/tools/server/webui/src
Pascal d0fa2c9fbb
Send reasoning content back to the model across turns via the reasoning_content API field (#21036)
* webui: send reasoning_content back to model in context

Preserve assistant reasoning across turns by extracting it from
internal tags and sending it as a separate reasoning_content field
in the API payload. The server and Jinja templates handle native
formatting (e.g. <think> tags for Qwen, GLM, DeepSeek...).

Adds "Exclude reasoning from context" toggle in Settings > Developer
(off by default, so reasoning is preserved). Includes unit tests.

* webui: add syncable parameter for excludeReasoningFromContext

* chore: update webui build output
2026-03-27 08:17:35 +01:00
..
lib Send reasoning content back to the model across turns via the reasoning_content API field (#21036) 2026-03-27 08:17:35 +01:00
routes webui: Fix duplicated messages on q param (#20715) 2026-03-18 10:32:43 +01:00
styles feat(webui): improve LaTeX rendering with currency detection (#16508) 2025-11-03 00:41:08 +01:00
app.css (webui) REFACTOR: UI primitives and polish (#19551) 2026-02-12 12:21:00 +01:00
app.d.ts webui: Fix selecting generated output issues during active streaming (#18091) 2025-12-18 11:13:52 +01:00
app.html SvelteKit-based WebUI (#14839) 2025-09-17 19:29:13 +02:00