SimpleChatTC:SettingsDefault:Enable cache prompt api option

This commit is contained in:
hanishkvc 2025-11-03 18:57:29 +05:30
parent 2cdf3f574c
commit 7fce3eeb2a
2 changed files with 15 additions and 6 deletions

View File

@ -583,21 +583,30 @@ Initial go at a pdftext tool call. It allows web / local pdf files to be read an
extracted and passed to ai model for further processing, as decided by ai and end user. One could extracted and passed to ai model for further processing, as decided by ai and end user. One could
either work with the full pdf or a subset of adjacent pages. either work with the full pdf or a subset of adjacent pages.
SimpleProxy SimpleProxy updates
* Convert from a single monolithic file into a collection of modules. * Convert from a single monolithic file into a collection of modules.
* UrlValidator to cross check scheme and domain of requested urls, * UrlValidator to cross check scheme and domain of requested urls,
the whitelist inturn picked from config json the whitelist inturn picked from config json
* Helpers to fetch file from local file system or the web, transparently * Helpers to fetch file from local file system or the web, transparently
* Help check for needed modules before a particular service path is acknowledged as available
through /aum service path
Chances are for ai models which dont support tool calling, things will be such that the tool calls Settings/Config default changes
meta data shared will be silently ignored without much issue. So enabling tool calling feature by default,
so that in case one is using a ai model with tool calling the feature is readily available for use.
Revert SlidingWindow ChatHistory in Context from last 10 to last 5 (rather 2 more then origianl, * Chances are for ai models which dont support tool calling, things will be such that the tool calls
meta data shared will be silently ignored without much issue. So enabling tool calling feature by
default, so that in case one is using a ai model with tool calling the feature is readily available
for use.
* Revert SlidingWindow ChatHistory in Context from last 10 to last 5 (rather 2 more then origianl,
given more context support in todays models) by default, given that now tool handshakes go through given more context support in todays models) by default, given that now tool handshakes go through
the tools related side channel in the http handshake and arent morphed into normal user-assistant the tools related side channel in the http handshake and arent morphed into normal user-assistant
channel of the handshake. channel of the handshake.
* Enable CachePrompt api option given that tool calling based interactions could involve chat sessions
having ai responses built over multiple steps of tool callings etal. So independent of our client side
sliding window based drop off or even before they kick in, this can help in many cases.
#### ToDo #### ToDo

View File

@ -1356,7 +1356,7 @@ class Me {
"temperature": 0.7, "temperature": 0.7,
"max_tokens": 2048, "max_tokens": 2048,
"n_predict": 2048, "n_predict": 2048,
"cache_prompt": false, "cache_prompt": true,
//"frequency_penalty": 1.2, //"frequency_penalty": 1.2,
//"presence_penalty": 1.2, //"presence_penalty": 1.2,
}; };