llama.cpp/tools/server/webui/src
Pascal 6ce3d85796
server: (webui) add --webui-config (#18028)
* server/webui: add server-side WebUI config support

Add CLI arguments --webui-config (inline JSON) and --webui-config-file
(file path) to configure WebUI default settings from server side.

Backend changes:
- Parse JSON once in server_context::load_model() for performance
- Cache parsed config in webui_settings member (zero overhead on /props)
- Add proper error handling in router mode with try/catch
- Expose webui_settings in /props endpoint for both router and child modes

Frontend changes:
- Add 14 configurable WebUI settings via parameter sync
- Add tests for webui settings extraction
- Fix subpath support with base path in API calls

Addresses feedback from @ngxson and @ggerganov

* server: address review feedback from ngxson

* server: regenerate README with llama-gen-docs
2025-12-17 21:45:45 +01:00
..
lib server: (webui) add --webui-config (#18028) 2025-12-17 21:45:45 +01:00
routes server: (webui) add --webui-config (#18028) 2025-12-17 21:45:45 +01:00
styles feat(webui): improve LaTeX rendering with currency detection (#16508) 2025-11-03 00:41:08 +01:00
app.css server: introduce API for serving / loading / unloading multiple models (#17470) 2025-12-01 19:41:04 +01:00
app.d.ts server: introduce API for serving / loading / unloading multiple models (#17470) 2025-12-01 19:41:04 +01:00
app.html SvelteKit-based WebUI (#14839) 2025-09-17 19:29:13 +02:00