llama.cpp/requirements
손희준 fbbf3ad190
server: /v1/responses (partial) (#18486)
* from previous PR

* Make instruction(system) as first message

* Convert [input_message] (text/image/file)

* Rename convert_responses_to_chatcmpl(body) -> response_body

* Initial tool call support

* Erase instructions field from chatcmpl body

* Feed reasoning texts to chat template

* Use std::vector instead of opaque json array

* Make output_item.added events consistent

* Move `server_task_result_cmpl_partial::update` from header to source

* Match ID of output_item.added and .done events

* Add function_call only if there is no "fc_" prefix

* Add function call output at non-streaming API

* Test if ID is persistent

* Add doc

* Fix style - use trailing comma

* Rewrite state management

* catch up with upstream/master

* Fix style - "type" is the first item of SSE data

* Explicitly check "instructions" from response_body

* Make lambdas static

* Check if reasoning content exists

* Add `oai_resp_id` to task_result_state(also initialized at ctor), server_task_result_cmpl_partial, and server_task_result_cmpl_final

* Reject `input_file` since it is not supported by chatcmpl

* Add "fc_" prefix to non-straming function call id as coderabbit pointed out

---------

Co-authored-by: openingnow <>
2026-01-21 17:47:23 +01:00
..
requirements-all.txt model-conversion : add support for SentenceTransformers (#16387) 2025-10-09 14:35:22 +02:00
requirements-compare-llama-bench.txt compare-llama-bench: add option to plot (#14169) 2025-06-14 10:34:20 +02:00
requirements-convert_hf_to_gguf.txt convert : Make mistral-common dependency optional (#16738) 2025-10-23 15:54:46 +02:00
requirements-convert_hf_to_gguf_update.txt ci : check that pre-tokenizer hashes are up-to-date (#15032) 2025-08-02 14:39:01 +02:00
requirements-convert_legacy_llama.txt convert : update transformers requirements (#16866) 2025-10-30 23:15:03 +01:00
requirements-convert_llama_ggml_to_gguf.txt py : switch to snake_case (#8305) 2024-07-05 07:53:33 +03:00
requirements-convert_lora_to_gguf.txt common: Include torch package for s390x (#13699) 2025-05-22 21:31:29 +03:00
requirements-gguf_editor_gui.txt gguf-py : add support for sub_type (in arrays) in GGUFWriter add_key_value method (#13561) 2025-05-29 15:36:05 +02:00
requirements-pydantic.txt mtmd : add support for Voxtral (#14862) 2025-07-28 15:01:48 +02:00
requirements-server-bench.txt scripts: benchmark for HTTP server throughput (#14668) 2025-07-14 13:14:30 +02:00
requirements-test-tokenizer-random.txt py : type-check all Python scripts with Pyright (#8341) 2024-07-07 15:04:39 -04:00
requirements-tool_bench.txt server: /v1/responses (partial) (#18486) 2026-01-21 17:47:23 +01:00