llama.cpp/tools/server/webui/src
Rohan Jain 974c8c94cc
webui: add setting for first-line chat titles (#21797)
* webui: add setting for first-line chat titles

Add an opt-in setting (`titleGenerationUseFirstLine`) to use the first
non-empty line of a prompt as the generated conversation title.

Previously, the complete multi-line prompt was being used, which created
long titles for complex queries. Coupled with
"Ask for confirmation before changing conversation title", the dialog
would overflow.

* Update tools/server/webui/src/lib/utils/text.ts

Co-authored-by: Aleksander Grygier <aleksander.grygier@gmail.com>

* Update tools/server/webui/src/lib/utils/text.ts

Co-authored-by: Aleksander Grygier <aleksander.grygier@gmail.com>

* webui: Run build to update the bundle

As requested in:
https://github.com/ggml-org/llama.cpp/pull/21797#pullrequestreview-4094935065

* webui: Fix missing import for NEWLINE_SEPARATOR

---------

Co-authored-by: Aleksander Grygier <aleksander.grygier@gmail.com>
2026-04-13 09:30:46 +02:00
..
lib webui: add setting for first-line chat titles (#21797) 2026-04-13 09:30:46 +02:00
routes fix: Model Selector choice sync (#21628) 2026-04-09 09:46:27 +02:00
styles feat(webui): improve LaTeX rendering with currency detection (#16508) 2025-11-03 00:41:08 +01:00
app.css server: (webui) no more gzip compression (#21073) 2026-03-31 15:44:26 +02:00
app.d.ts webui: Fix selecting generated output issues during active streaming (#18091) 2025-12-18 11:13:52 +01:00
app.html SvelteKit-based WebUI (#14839) 2025-09-17 19:29:13 +02:00