llama.cpp/tools
Aleksander Grygier 1a18927894
Allow viewing conversations even when llama server is down (#16255)
* webui: allow viewing conversations and sending messages even if llama-server is down

- Cached llama.cpp server properties in browser localStorage on startup, persisting successful fetches and reloading them when refresh attempts fail so the chat UI continues to render while the backend is unavailable.
- Cleared the stored server properties when resetting the store to prevent stale capability data after cache-backed operation.
- Kept the original error-splash behavior when no cached props exist so fresh installs still surface a clear failure state instead of rendering stale data.

* feat: Add UI for `props` endpoint unavailable + cleanup logic

* webui: extend cached props fallback to offline errors

Treat connection failures (refused, DNS, timeout, fetch) the same way as
server 5xx so the warning banner shows up when cache is available, instead
of falling back to a full error screen.

* webui: Left the chat form enabled when a server warning is present so operators can keep sending messages

e.g., to restart the backend over llama-swap, even while cached /props data is in use

* chore: update webui build output

---------

Co-authored-by: Pascal <admin@serveurperso.com>
2025-09-26 18:35:42 +02:00
..
batched-bench cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
cvector-generator cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
export-lora cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
gguf-split ci : use smaller model (#16168) 2025-09-22 09:11:39 +03:00
imatrix cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
llama-bench llama-bench: add --devices and --list-devices support (#16039) 2025-09-20 00:15:21 +02:00
main ci : adjust params for less runtime (#16167) 2025-09-22 08:31:40 +03:00
mtmd mtmd : fix uninitialized variable in bicubic_resize (#16275) 2025-09-26 15:00:44 +02:00
perplexity llama: print memory breakdown on exit (#15860) 2025-09-24 16:53:48 +02:00
quantize ci : use smaller model (#16168) 2025-09-22 09:11:39 +03:00
rpc rpc : fix regression when --device is used (#15981) 2025-09-14 12:28:18 +03:00
run cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
server Allow viewing conversations even when llama server is down (#16255) 2025-09-26 18:35:42 +02:00
tokenize cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
tts cmake : Do not install tools on iOS targets (#15903) 2025-09-16 09:54:44 +07:00
CMakeLists.txt mtmd : rename llava directory to mtmd (#13311) 2025-05-05 16:02:55 +02:00