2421 lines
95 KiB
C++
2421 lines
95 KiB
C++
// Tests chat handling, including grammar generation and parsing for tool calling, for various templates.
|
||
//
|
||
// Also acts as a CLI to generate a Markdown summary of the formats of Jinja templates,
|
||
// e.g. given Minja (http://github.com/google/minja) checked out in parent dir:
|
||
//
|
||
// cmake -B build && cmake --build build --parallel && ./build/bin/test-chat ../minja/build/tests/*.jinja 2>/dev/null
|
||
//
|
||
#include "../src/llama-grammar.h"
|
||
#include "../src/unicode.h"
|
||
#include "chat-auto-parser.h"
|
||
#include "chat.h"
|
||
#include "common.h"
|
||
#include "ggml.h"
|
||
#include "log.h"
|
||
|
||
#include <algorithm>
|
||
#include <exception>
|
||
#include <fstream>
|
||
#include <functional>
|
||
#include <iostream>
|
||
#include <nlohmann/json.hpp>
|
||
#include <stdexcept>
|
||
#include <string>
|
||
|
||
using json = nlohmann::ordered_json;
|
||
|
||
static std::ostream & operator<<(std::ostream & os, const common_chat_msg_diff & diff) {
|
||
os << "{ content_delta: " << diff.content_delta << "; ";
|
||
os << "reasoning_content_delta: " << diff.reasoning_content_delta << "; ";
|
||
if (diff.tool_call_index != std::string::npos) {
|
||
os << "tool_call_index: " << diff.tool_call_index << "; ";
|
||
os << "tool_call_delta.name: " << diff.tool_call_delta.name << "; ";
|
||
os << "tool_call_delta.id: " << diff.tool_call_delta.id << "; ";
|
||
os << "tool_call_delta.arguments: " << diff.tool_call_delta.arguments << "; ";
|
||
}
|
||
os << "}";
|
||
return os;
|
||
}
|
||
|
||
// operator<< for vector<common_chat_msg_diff>:
|
||
static std::ostream & operator<<(std::ostream & os, const std::vector<common_chat_msg_diff> & diffs) {
|
||
os << "[\n";
|
||
for (const auto & diff : diffs) {
|
||
os << " " << diff << ",\n";
|
||
}
|
||
os << "]";
|
||
return os;
|
||
}
|
||
|
||
static std::ostream & operator<<(std::ostream & os, const common_chat_msg & msg) {
|
||
os << "{ role: " << msg.role << "; ";
|
||
os << "content: " << msg.content << "; ";
|
||
os << "content_parts: [\n";
|
||
for (const auto & part : msg.content_parts) {
|
||
os << " { type: " << part.type << "; text: " << part.text << " },\n";
|
||
}
|
||
os << "]; ";
|
||
os << "reasoning_content: " << msg.reasoning_content << "; ";
|
||
os << "tool_calls: [\n";
|
||
for (const auto & tool_call : msg.tool_calls) {
|
||
os << " { name: " << tool_call.name << "; arguments: " << tool_call.arguments << "; id: " << tool_call.id
|
||
<< " },\n";
|
||
}
|
||
os << "]";
|
||
os << "}";
|
||
return os;
|
||
}
|
||
|
||
template <class T> static bool equals(const T & expected, const T & actual) {
|
||
return expected == actual;
|
||
}
|
||
|
||
static common_chat_msg normalize(const common_chat_msg & msg) {
|
||
common_chat_msg normalized = msg;
|
||
for (auto & tool_call : normalized.tool_calls) {
|
||
try {
|
||
tool_call.arguments = json::parse(tool_call.arguments).dump();
|
||
} catch (const std::exception &) {
|
||
}
|
||
}
|
||
return normalized;
|
||
}
|
||
|
||
template <> bool equals(const common_chat_msg & expected, const common_chat_msg & actual) {
|
||
return normalize(expected) == normalize(actual);
|
||
}
|
||
|
||
template <class T> static void assert_equals(const T & expected, const T & actual) {
|
||
if (!equals(expected, actual)) {
|
||
std::ostringstream oss_expected;
|
||
oss_expected << expected;
|
||
std::ostringstream oss_actual;
|
||
oss_actual << actual;
|
||
LOG_ERR("Expected: %s\n", oss_expected.str().c_str());
|
||
LOG_ERR("Actual: %s\n", oss_actual.str().c_str());
|
||
common_log_flush(common_log_main());
|
||
throw std::runtime_error("Test failed");
|
||
}
|
||
}
|
||
|
||
static std::string read_file(const std::string & path) {
|
||
std::ifstream fs(path, std::ios_base::binary);
|
||
if (!fs.is_open()) {
|
||
fs = std::ifstream("../" + path, std::ios_base::binary);
|
||
if (!fs.is_open()) {
|
||
throw std::runtime_error("Failed to open file: " + path);
|
||
}
|
||
}
|
||
fs.seekg(0, std::ios_base::end);
|
||
auto size = fs.tellg();
|
||
fs.seekg(0);
|
||
std::string out;
|
||
out.resize(static_cast<size_t>(size));
|
||
fs.read(out.data(), static_cast<std::streamsize>(size));
|
||
return out;
|
||
}
|
||
|
||
static common_chat_templates_ptr read_templates(const std::string & path) {
|
||
return common_chat_templates_ptr(common_chat_templates_init(/* model= */ nullptr, read_file(path)));
|
||
}
|
||
|
||
static std::unique_ptr<llama_grammar> build_grammar(const std::string & grammar_str) {
|
||
return std::unique_ptr<llama_grammar>(
|
||
llama_grammar_init_impl(nullptr, grammar_str.c_str(), "root", false, nullptr, 0, nullptr, 0));
|
||
}
|
||
|
||
// TODO: extract to common helper (copied from test-grammar-integration.cpp)
|
||
static bool match_string(const std::string & input, llama_grammar * grammar) {
|
||
const auto cpts = unicode_cpts_from_utf8(input);
|
||
|
||
auto & stacks_cur = llama_grammar_get_stacks(grammar);
|
||
|
||
for (const auto & cpt : cpts) {
|
||
llama_grammar_accept(grammar, cpt);
|
||
|
||
if (stacks_cur.empty()) {
|
||
// no stacks means that the grammar failed to match at this point
|
||
return false;
|
||
}
|
||
}
|
||
|
||
if (std::any_of(stacks_cur.begin(), stacks_cur.end(), [](const auto & stack) { return stack.empty(); })) {
|
||
// An empty stack means that the grammar has been completed
|
||
return true;
|
||
}
|
||
|
||
return false;
|
||
}
|
||
|
||
static std::string renormalize_json(const std::string & json_str) {
|
||
try {
|
||
auto json_obj = json::parse(json_str);
|
||
return json_obj.dump();
|
||
} catch (const std::exception & e) {
|
||
return ""; // ignore parial JSON contents for comparison purposes
|
||
}
|
||
}
|
||
|
||
static void assert_msg_equals(const common_chat_msg & expected,
|
||
const common_chat_msg & actual,
|
||
bool ignore_whitespace_differences = false) {
|
||
assert_equals(expected.role, actual.role);
|
||
if (ignore_whitespace_differences) {
|
||
assert_equals(string_strip(expected.content), string_strip(actual.content));
|
||
} else {
|
||
assert_equals(expected.content, actual.content);
|
||
}
|
||
assert_equals(expected.content_parts.size(), actual.content_parts.size());
|
||
for (size_t i = 0; i < expected.content_parts.size(); i++) {
|
||
const auto & expected_part = expected.content_parts[i];
|
||
const auto & actual_part = actual.content_parts[i];
|
||
assert_equals(expected_part.type, actual_part.type);
|
||
if (ignore_whitespace_differences) {
|
||
assert_equals(string_strip(expected_part.text), string_strip(actual_part.text));
|
||
} else {
|
||
assert_equals(expected_part.text, actual_part.text);
|
||
}
|
||
}
|
||
if (ignore_whitespace_differences) {
|
||
assert_equals(string_strip(expected.reasoning_content), string_strip(actual.reasoning_content));
|
||
} else {
|
||
assert_equals(expected.reasoning_content, actual.reasoning_content);
|
||
}
|
||
assert_equals(expected.tool_calls.size(), actual.tool_calls.size());
|
||
for (size_t i = 0; i < expected.tool_calls.size(); i++) {
|
||
const auto & expected_tool_call = expected.tool_calls[i];
|
||
const auto & actual_tool_call = actual.tool_calls[i];
|
||
assert_equals(expected_tool_call.name, actual_tool_call.name);
|
||
assert_equals(renormalize_json(expected_tool_call.arguments), renormalize_json(actual_tool_call.arguments));
|
||
assert_equals(expected_tool_call.id, actual_tool_call.id);
|
||
}
|
||
}
|
||
|
||
static common_chat_tool special_function_tool{
|
||
/* .name = */ "special_function",
|
||
/* .description = */ "I'm special",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"arg1": {
|
||
"type": "integer",
|
||
"description": "The arg."
|
||
}
|
||
},
|
||
"required": ["arg1"]
|
||
})",
|
||
};
|
||
static common_chat_tool special_function_tool_with_optional_param{
|
||
/* .name = */ "special_function_with_opt",
|
||
/* .description = */ "I'm special but have optional stuff",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"arg1": {
|
||
"type": "integer",
|
||
"description": "The arg."
|
||
},
|
||
"arg2": {
|
||
"type": "integer",
|
||
"description": "The optional arg."
|
||
}
|
||
},
|
||
"required": ["arg1"]
|
||
})",
|
||
};
|
||
static common_chat_tool python_tool{
|
||
/* .name = */ "python",
|
||
/* .description = */ "an ipython interpreter",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"code": {
|
||
"type": "string",
|
||
"description": "Python code to execute."
|
||
}
|
||
},
|
||
"required": ["code"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool html_tool{
|
||
/* .name = */ "html",
|
||
/* .description = */ "an html validator",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"markup": {
|
||
"type": "string",
|
||
"description": "HTML markup to validate."
|
||
}
|
||
},
|
||
"required": ["markup"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool get_time_tool{
|
||
/* .name = */ "get_time",
|
||
/* .description = */ "Get the current time in a city",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"city": {
|
||
"type": "string",
|
||
"description": "City name"
|
||
}
|
||
},
|
||
"required": ["city"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool get_weather_tool{
|
||
/* .name = */ "get_weather",
|
||
/* .description = */ "Get the current weather in a city",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"city": {
|
||
"type": "string",
|
||
"description": "City name"
|
||
}
|
||
},
|
||
"required": ["city"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool todo_list{
|
||
/* .name = */ "todo_list",
|
||
/* .description = */ "Create or update the todo list",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"todos": {
|
||
"type": "array",
|
||
"description": "List of TODO list items"
|
||
}
|
||
},
|
||
"required": ["todos"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool edit_tool{
|
||
/* .name = */ "edit",
|
||
/* .description = */ "Edit file",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"filename": {
|
||
"type": "string",
|
||
"description": "Path of file to edit"
|
||
},
|
||
"oldString": {
|
||
"type": "string",
|
||
"description": "String to replace"
|
||
},
|
||
"newString": {
|
||
"type": "string",
|
||
"description": "New (replacement) value"
|
||
}
|
||
},
|
||
"required": ["filename", "oldString", "newString"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool magic_tool{
|
||
/* .name = */ "magic",
|
||
/* .description = */ "Magic tool that takes a hash",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"name": {
|
||
"type": "string"
|
||
},
|
||
"ref": {
|
||
"type": "string"
|
||
}
|
||
},
|
||
"required": ["name", "ref"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool magic_int_tool{
|
||
/* .name = */ "magic_int",
|
||
/* .description = */ "Magic tool that takes a hash",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"ref": {
|
||
"type": "integer"
|
||
},
|
||
"name": {
|
||
"type": "string"
|
||
}
|
||
},
|
||
"required": ["ref"]
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool string_param_tool{
|
||
/* .name = */ "string_param",
|
||
/* .description = */ "Tool with string parameter for testing",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"text": {
|
||
"type": "string",
|
||
"description": "A text parameter"
|
||
}
|
||
},
|
||
"required": []
|
||
})",
|
||
};
|
||
|
||
static common_chat_tool quoted_unquoted_tool{
|
||
/* .name = */ "quoted_unquoted",
|
||
/* .description = */ "Tool with two string parameters, one for quoted string, one for unquoted",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"quoted": {
|
||
"type": "string",
|
||
"description": "Quoted value"
|
||
},
|
||
"unquoted": {
|
||
"type": "string",
|
||
"description": "Unquoted value"
|
||
}
|
||
},
|
||
"required": ["quoted", "unquoted"]
|
||
})",
|
||
};
|
||
|
||
|
||
static std::vector<common_chat_tool> tools{ special_function_tool, special_function_tool_with_optional_param,
|
||
python_tool, html_tool, todo_list };
|
||
|
||
const common_chat_msg message_user{
|
||
"user",
|
||
"Hey there!",
|
||
/* .content_parts = */ {},
|
||
/* .tool_calls = */ {},
|
||
/* .reasoning_content = */ "",
|
||
/* .tool_name = */ "",
|
||
/* .tool_call_id = */ "",
|
||
};
|
||
|
||
const common_chat_msg message_user_parts{
|
||
"user",
|
||
/* .content = */ "",
|
||
/* .content_parts = */
|
||
{
|
||
{ "text", "Hey" },
|
||
{ "text", "there" },
|
||
},
|
||
/* .tool_calls = */
|
||
{ },
|
||
/* .reasoning_content = */
|
||
"",
|
||
/* .tool_name = */ "",
|
||
/* .tool_call_id = */ "",
|
||
};
|
||
|
||
static common_chat_msg simple_assist_msg(const std::string & content,
|
||
const std::string & reasoning_content = "",
|
||
const std::string & tool_name = "",
|
||
const std::string & arguments = "",
|
||
const std::string & id = "") {
|
||
common_chat_msg msg;
|
||
msg.role = "assistant";
|
||
msg.content = content;
|
||
msg.reasoning_content = reasoning_content;
|
||
if (!tool_name.empty() || !id.empty()) {
|
||
msg.tool_calls.push_back({ tool_name, arguments, id });
|
||
}
|
||
return msg;
|
||
}
|
||
|
||
static common_chat_msg message_with_tool_calls(const std::string & tool_name, const std::string & arguments) {
|
||
return simple_assist_msg("", "", tool_name, arguments);
|
||
}
|
||
|
||
static common_chat_msg message_with_tool_calls_and_reasoning(const std::string & tool_name,
|
||
const std::string & arguments,
|
||
const std::string & reasoning) {
|
||
return simple_assist_msg("", reasoning, tool_name, arguments);
|
||
}
|
||
|
||
static common_chat_msg message_with_reasoning_content_and_multiple_tool_calls(
|
||
const std::string & reasoning,
|
||
const std::string & content,
|
||
const std::vector<std::pair<std::string, std::string>> & tool_calls) {
|
||
common_chat_msg msg;
|
||
msg.role = "assistant";
|
||
msg.content = content;
|
||
msg.reasoning_content = reasoning;
|
||
for (const auto & [name, args] : tool_calls) {
|
||
msg.tool_calls.push_back({ name, args, "" });
|
||
}
|
||
return msg;
|
||
}
|
||
|
||
static common_chat_msg message_with_content_and_tool_call(const std::string & content,
|
||
const std::string & tool_name,
|
||
const std::string & arguments) {
|
||
return simple_assist_msg(content, "", tool_name, arguments);
|
||
}
|
||
|
||
static common_chat_msg message_with_reasoning_and_tool_call(const std::string & reasoning,
|
||
const std::string & tool_name,
|
||
const std::string & arguments) {
|
||
return simple_assist_msg("", reasoning, tool_name, arguments);
|
||
}
|
||
|
||
const common_chat_msg message_assist = simple_assist_msg("Hello, world!\nWhat's up?");
|
||
const common_chat_msg message_assist_empty = simple_assist_msg("");
|
||
const common_chat_msg message_assist_thoughts_unparsed_deepseek =
|
||
simple_assist_msg("<think>I'm\nthinking</think>Hello, world!\nWhat's up?");
|
||
const common_chat_msg message_assist_thoughts_unparsed_md =
|
||
simple_assist_msg("<think>I'm\nthinking</think>Hello, world!\nWhat's up?\n```json\n{}```");
|
||
const common_chat_msg message_assist_thoughts_unparsed_md_partial =
|
||
simple_assist_msg("<think>I'm\nthinking</think>Hello, world!\nWhat's up?\n```json\n{}");
|
||
|
||
const common_chat_msg message_assist_thoughts_unparsed_r7b =
|
||
simple_assist_msg("<|START_THINKING|>I'm\nthinking<|END_THINKING|>Hello, world!\nWhat's up?");
|
||
const common_chat_msg message_assist_thoughts_unparsed_magistral =
|
||
simple_assist_msg("[THINK]raisonnement[/THINK]Réponse");
|
||
const common_chat_msg message_assist_thoughts = simple_assist_msg("Hello, world!\nWhat's up?", "I'm\nthinking");
|
||
const common_chat_msg message_assist_thoughts_unopened_unparsed =
|
||
simple_assist_msg("I'm\nthinking</think>Hello, world!\nWhat's up?");
|
||
const common_chat_msg message_assist_thoughts_no_content = simple_assist_msg("", "I'm\nthinking");
|
||
const common_chat_msg message_assist_call = simple_assist_msg("", "", "special_function", "{\"arg1\": 1}");
|
||
const common_chat_msg message_assist_call_noopt =
|
||
simple_assist_msg("", "", "special_function_with_opt", "{\"arg1\": 1}");
|
||
const common_chat_msg message_assist_call_withopt =
|
||
simple_assist_msg("", "", "special_function_with_opt", "{\"arg1\": 1, \"arg2\": 2}");
|
||
const common_chat_msg message_assist_call_content =
|
||
simple_assist_msg("Hello, world!\nWhat's up?", "", "special_function", "{\"arg1\":1}");
|
||
const common_chat_msg message_assist_call_empty_args = simple_assist_msg("", "", "special_function");
|
||
const common_chat_msg message_assist_call_cutoff_args = simple_assist_msg("", "", "special_function", "{\"arg");
|
||
const common_chat_msg message_assist_call_thoughts =
|
||
simple_assist_msg("", "I'm\nthinking", "special_function", "{\"arg1\":1}");
|
||
const common_chat_msg message_assist_call_thoughts_unparsed =
|
||
simple_assist_msg("<think>I'm\nthinking</think>\n\n", "", "special_function", "{\"arg1\": 1}");
|
||
const common_chat_msg message_assist_call_thoughts_content =
|
||
simple_assist_msg("Hello, world!\nWhat's up?", "I'm\nthinking", "special_function", "{\"arg1\": 1}");
|
||
const common_chat_msg message_assist_call_id =
|
||
simple_assist_msg("", "", "special_function", "{\"arg1\":1}", /* .id = */ "123456789");
|
||
const common_chat_msg message_assist_call_idx =
|
||
simple_assist_msg("", "", "special_function", "{\"arg1\":1}", /* .id = */ "0");
|
||
const common_chat_msg message_assist_thoughts_call_idx =
|
||
simple_assist_msg("", "I'm\nthinking", "special_function", "{\"arg1\": 1}", /* id = */ "0");
|
||
const common_chat_msg message_assist_thoughts_partial_call =
|
||
simple_assist_msg("", "I'm\nthinking", "special_function", "", /* id = */ "0");
|
||
const common_chat_msg message_assist_call_python = simple_assist_msg("", "", "python", "{\"code\":\"print('hey')\"}");
|
||
const common_chat_msg message_assist_call_python_lines =
|
||
simple_assist_msg("", "", "python", "{\"code\":\"# This is a program:\\nprint('hey')\"}");
|
||
const common_chat_msg message_assist_call_python_lines_unclosed =
|
||
simple_assist_msg("", "", "python", "{\"code\":\"# This is a program:\\nprint('hey')");
|
||
const common_chat_msg message_assist_json_content =
|
||
simple_assist_msg("{\n \"response\": \"Hello, world!\\nWhat's up?\"\n}");
|
||
|
||
struct delta_data {
|
||
std::string delta;
|
||
common_chat_params params;
|
||
};
|
||
|
||
static delta_data init_delta(const struct common_chat_templates * tmpls,
|
||
const std::vector<std::string> & end_tokens,
|
||
const common_chat_msg & user_message,
|
||
const common_chat_msg & delta_message,
|
||
const std::vector<common_chat_tool> & tools,
|
||
const common_chat_tool_choice & tool_choice) {
|
||
common_chat_templates_inputs inputs;
|
||
inputs.parallel_tool_calls = true;
|
||
inputs.messages.push_back(user_message);
|
||
inputs.tools = tools;
|
||
inputs.tool_choice = tool_choice;
|
||
auto params_prefix = common_chat_templates_apply(tmpls, inputs);
|
||
|
||
inputs.messages.push_back(delta_message);
|
||
inputs.add_generation_prompt = false;
|
||
auto params_full = common_chat_templates_apply(tmpls, inputs);
|
||
|
||
std::string prefix = params_prefix.prompt;
|
||
std::string full = params_full.prompt;
|
||
|
||
if (full == prefix) {
|
||
throw std::runtime_error("Full message is the same as the prefix");
|
||
}
|
||
|
||
size_t common_prefix_length = 0;
|
||
for (size_t i = 0; i < prefix.size() && i < full.size(); ++i) {
|
||
if (prefix[i] != full[i]) {
|
||
break;
|
||
}
|
||
if (prefix[i] == '<') {
|
||
// DeepSeek R1's template (as of 20250209) adds a trailing <think> if add_generation_prompt,
|
||
// but it removes thinking tags for past messages.
|
||
// The prefix and full strings diverge at <think> vs. <|tool▁calls▁begin|>, we avoid consuming the leading <.
|
||
continue;
|
||
}
|
||
common_prefix_length = i + 1;
|
||
}
|
||
auto delta = full.substr(common_prefix_length);
|
||
|
||
// Strip end tokens
|
||
for (const auto & end_token : end_tokens) {
|
||
// rfind to find the last occurrence
|
||
auto pos = delta.rfind(end_token);
|
||
if (pos != std::string::npos) {
|
||
delta = delta.substr(0, pos);
|
||
break;
|
||
}
|
||
}
|
||
return { delta, params_full };
|
||
}
|
||
|
||
/*
|
||
Applies the template to 1 user message w/ add_generation_prompt=true, then w/ the test message w/ add_generation_prompt=false,
|
||
gets the diff, removes any end tokens and parses the result w/ the grammar, checking that
|
||
the parsed message is the same as the test_message
|
||
*/
|
||
static void test_templates(const struct common_chat_templates * tmpls,
|
||
const std::vector<std::string> & end_tokens,
|
||
const common_chat_msg & test_message,
|
||
const std::vector<common_chat_tool> & tools = {},
|
||
const std::string & expected_delta = "",
|
||
bool expect_grammar_triggered = true,
|
||
bool test_grammar_if_triggered = true,
|
||
common_reasoning_format reasoning_format = COMMON_REASONING_FORMAT_NONE,
|
||
bool ignore_whitespace_differences = false) {
|
||
common_chat_msg user_message;
|
||
user_message.role = "user";
|
||
user_message.content = "Hello, world!";
|
||
|
||
common_chat_templates_inputs inputs_tools;
|
||
inputs_tools.messages = { message_user };
|
||
inputs_tools.tools = { special_function_tool };
|
||
|
||
common_chat_params params = common_chat_templates_apply(tmpls, inputs_tools);
|
||
|
||
for (const auto & tool_choice :
|
||
std::vector<common_chat_tool_choice>{ COMMON_CHAT_TOOL_CHOICE_AUTO, COMMON_CHAT_TOOL_CHOICE_REQUIRED }) {
|
||
auto data = init_delta(tmpls, end_tokens, user_message, test_message, tools, tool_choice);
|
||
if (!expected_delta.empty()) {
|
||
if (ignore_whitespace_differences) {
|
||
assert_equals(string_strip(expected_delta), string_strip(data.delta));
|
||
} else {
|
||
assert_equals(expected_delta, data.delta);
|
||
}
|
||
}
|
||
|
||
if (expect_grammar_triggered) {
|
||
// TODO @ngxson : refactor common_chat_parse to avoid passing format/reasoning_format every time
|
||
common_chat_parser_params parser_params;
|
||
parser_params.format = data.params.format;
|
||
parser_params.reasoning_format = reasoning_format;
|
||
if (!parser_params.parser.empty()) {
|
||
parser_params.parser = common_peg_arena();
|
||
parser_params.parser.load(params.parser);
|
||
}
|
||
const auto msg = common_chat_parse(data.delta, /* is_partial= */ false, parser_params);
|
||
assert_msg_equals(test_message, msg, ignore_whitespace_differences);
|
||
}
|
||
|
||
if (!test_message.tool_calls.empty()) {
|
||
GGML_ASSERT(!data.params.grammar.empty());
|
||
}
|
||
if (!data.params.grammar.empty()) {
|
||
auto grammar = build_grammar(data.params.grammar);
|
||
if (!grammar) {
|
||
throw std::runtime_error("Failed to build grammar");
|
||
}
|
||
auto earliest_trigger_pos = std::string::npos;
|
||
auto constrained = data.delta;
|
||
for (const auto & trigger : data.params.grammar_triggers) {
|
||
size_t pos = std::string::npos;
|
||
std::smatch match;
|
||
switch (trigger.type) {
|
||
case COMMON_GRAMMAR_TRIGGER_TYPE_WORD:
|
||
{
|
||
const auto & word = trigger.value;
|
||
pos = constrained.find(word);
|
||
break;
|
||
}
|
||
case COMMON_GRAMMAR_TRIGGER_TYPE_PATTERN:
|
||
{
|
||
const auto & pattern = trigger.value;
|
||
if (std::regex_search(constrained, match, std::regex(pattern))) {
|
||
pos = match.position(1);
|
||
}
|
||
break;
|
||
}
|
||
case COMMON_GRAMMAR_TRIGGER_TYPE_PATTERN_FULL:
|
||
{
|
||
const auto & pattern = trigger.value;
|
||
if (std::regex_match(constrained, match, std::regex(pattern))) {
|
||
auto mpos = std::string::npos;
|
||
for (size_t i = 1; i < match.size(); ++i) {
|
||
if (match[i].length() > 0) {
|
||
mpos = match.position(i);
|
||
break;
|
||
}
|
||
}
|
||
if (mpos == std::string::npos) {
|
||
mpos = match.position(0);
|
||
}
|
||
pos = mpos;
|
||
}
|
||
break;
|
||
}
|
||
default:
|
||
throw std::runtime_error("Unknown trigger type");
|
||
}
|
||
if (pos == std::string::npos) {
|
||
continue;
|
||
}
|
||
if (earliest_trigger_pos == std::string::npos || pos < earliest_trigger_pos) {
|
||
earliest_trigger_pos = pos;
|
||
}
|
||
}
|
||
auto grammar_triggered = false;
|
||
if (earliest_trigger_pos != std::string::npos) {
|
||
constrained = constrained.substr(earliest_trigger_pos);
|
||
grammar_triggered = true;
|
||
}
|
||
if (data.params.grammar_lazy) {
|
||
assert_equals(expect_grammar_triggered, grammar_triggered);
|
||
}
|
||
|
||
if (grammar_triggered && test_grammar_if_triggered && !match_string(constrained, grammar.get())) {
|
||
throw std::runtime_error("Failed to match delta against grammar:\n\n" + data.delta +
|
||
"\n\nConstrained: " + constrained + "\n\nGrammar: " + data.params.grammar);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Test if streaming=true is consistant with streaming=false for given partial parser
|
||
* Also test if there is any problem with partial message
|
||
*/
|
||
template <typename T>
|
||
static void test_parser_with_streaming(const common_chat_msg & expected, const std::string & raw_message, T parse_msg) {
|
||
constexpr auto utf8_truncate_safe_len = [](const std::string_view s) -> size_t {
|
||
auto len = s.size();
|
||
if (len == 0) {
|
||
return 0;
|
||
}
|
||
auto i = len;
|
||
for (size_t back = 0; back < 4 && i > 0; ++back) {
|
||
--i;
|
||
unsigned char c = s[i];
|
||
if ((c & 0x80) == 0) {
|
||
return len;
|
||
}
|
||
if ((c & 0xC0) == 0xC0) {
|
||
size_t expected_len = 0;
|
||
if ((c & 0xE0) == 0xC0) {
|
||
expected_len = 2;
|
||
} else if ((c & 0xF0) == 0xE0) {
|
||
expected_len = 3;
|
||
} else if ((c & 0xF8) == 0xF0) {
|
||
expected_len = 4;
|
||
} else {
|
||
return i;
|
||
}
|
||
if (len - i >= expected_len) {
|
||
return len;
|
||
}
|
||
return i;
|
||
}
|
||
}
|
||
return len - std::min(len, size_t(3));
|
||
};
|
||
constexpr auto utf8_truncate_safe_view = [utf8_truncate_safe_len](const std::string_view s) {
|
||
return s.substr(0, utf8_truncate_safe_len(s));
|
||
};
|
||
|
||
auto merged = simple_assist_msg("");
|
||
auto last_msg = parse_msg("");
|
||
for (size_t i = 1; i <= raw_message.size(); ++i) {
|
||
auto curr_msg = parse_msg(std::string(utf8_truncate_safe_view(std::string_view(raw_message).substr(0, i))));
|
||
if (curr_msg == simple_assist_msg("")) {
|
||
continue;
|
||
}
|
||
LOG_INF("Streaming msg: %s\n", common_chat_msgs_to_json_oaicompat({ curr_msg }).dump().c_str());
|
||
for (auto diff : common_chat_msg_diff::compute_diffs(last_msg, curr_msg)) {
|
||
LOG_INF("Streaming diff: %s\n", common_chat_msg_diff_to_json_oaicompat(diff).dump().c_str());
|
||
if (!diff.reasoning_content_delta.empty()) {
|
||
merged.reasoning_content += diff.reasoning_content_delta;
|
||
}
|
||
if (!diff.content_delta.empty()) {
|
||
merged.content += diff.content_delta;
|
||
}
|
||
if (diff.tool_call_index != std::string::npos) {
|
||
if (!diff.tool_call_delta.name.empty()) {
|
||
merged.tool_calls.push_back({ diff.tool_call_delta.name, "", "" });
|
||
}
|
||
if (!diff.tool_call_delta.arguments.empty()) {
|
||
GGML_ASSERT(!merged.tool_calls.empty());
|
||
merged.tool_calls.back().arguments += diff.tool_call_delta.arguments;
|
||
}
|
||
}
|
||
LOG_INF("Streaming merged: %s\n", common_chat_msgs_to_json_oaicompat({ merged }).dump().c_str());
|
||
}
|
||
assert_msg_equals(curr_msg, merged, true);
|
||
last_msg = curr_msg;
|
||
}
|
||
assert_msg_equals(expected, parse_msg(raw_message), true);
|
||
assert_msg_equals(expected, merged, true);
|
||
}
|
||
|
||
// Use for PEG parser implementations
|
||
struct peg_test_case {
|
||
common_chat_templates_inputs params;
|
||
std::string input;
|
||
common_chat_msg expect;
|
||
bool is_partial = false;
|
||
};
|
||
|
||
struct make_peg_parser {
|
||
common_chat_params params_;
|
||
common_peg_arena arena_;
|
||
bool detailed_debug_;
|
||
|
||
make_peg_parser(common_chat_templates * tmpls,
|
||
const common_chat_templates_inputs & inputs,
|
||
bool detailed_debug = false) {
|
||
detailed_debug_ = detailed_debug;
|
||
params_ = common_chat_templates_apply(tmpls, inputs);
|
||
arena_.load(params_.parser);
|
||
}
|
||
|
||
common_chat_msg parse(const std::string & msg, bool is_partial) const {
|
||
common_chat_parser_params parser_params;
|
||
parser_params.format = params_.format;
|
||
parser_params.debug = detailed_debug_;
|
||
return common_chat_peg_parse(arena_, msg, is_partial, parser_params);
|
||
}
|
||
};
|
||
|
||
static void test_peg_parser(common_chat_templates * tmpls,
|
||
const std::function<void(peg_test_case &)> & init,
|
||
bool detailed_debug) {
|
||
// UTF-8-safe truncation helper (same as in test_parser_with_streaming)
|
||
constexpr auto utf8_truncate_safe_len = [](const std::string_view s) -> size_t {
|
||
auto len = s.size();
|
||
if (len == 0) {
|
||
return 0;
|
||
}
|
||
auto i = len;
|
||
for (size_t back = 0; back < 4 && i > 0; ++back) {
|
||
--i;
|
||
unsigned char c = s[i];
|
||
if ((c & 0x80) == 0) {
|
||
return len;
|
||
}
|
||
if ((c & 0xC0) == 0xC0) {
|
||
size_t expected_len = 0;
|
||
if ((c & 0xE0) == 0xC0) {
|
||
expected_len = 2;
|
||
} else if ((c & 0xF0) == 0xE0) {
|
||
expected_len = 3;
|
||
} else if ((c & 0xF8) == 0xF0) {
|
||
expected_len = 4;
|
||
} else {
|
||
return i;
|
||
}
|
||
if (len - i >= expected_len) {
|
||
return len;
|
||
}
|
||
return i;
|
||
}
|
||
}
|
||
return len - std::min(len, size_t(3));
|
||
};
|
||
|
||
peg_test_case tc;
|
||
init(tc);
|
||
if (tc.params.messages.empty()) {
|
||
tc.params.messages = { message_user };
|
||
}
|
||
if (tc.expect.role.empty()) {
|
||
tc.expect.role = "assistant";
|
||
}
|
||
|
||
auto parser = make_peg_parser(tmpls, tc.params, detailed_debug);
|
||
if (detailed_debug) {
|
||
LOG_DBG("Using parser: \n%s\n", parser.arena_.dump(parser.arena_.root()).c_str());
|
||
}
|
||
|
||
common_chat_msg msg_accum;
|
||
common_chat_msg msg_prev;
|
||
msg_accum.role = msg_prev.role = "assistant";
|
||
|
||
for (size_t i = 1; i <= tc.input.size(); ++i) {
|
||
auto is_partial = i < tc.input.size() || tc.is_partial;
|
||
// Use UTF-8 safe truncation to avoid corrupting multi-byte characters
|
||
size_t safe_len = utf8_truncate_safe_len(std::string_view(tc.input).substr(0, i));
|
||
std::string prefix = tc.input.substr(0, safe_len);
|
||
common_chat_msg msg_current = parser.parse(prefix, is_partial);
|
||
|
||
for (const auto & diff : common_chat_msg_diff::compute_diffs(msg_prev, msg_current)) {
|
||
if (!diff.reasoning_content_delta.empty()) {
|
||
msg_accum.reasoning_content += diff.reasoning_content_delta;
|
||
}
|
||
if (!diff.content_delta.empty()) {
|
||
msg_accum.content += diff.content_delta;
|
||
}
|
||
if (diff.tool_call_index != std::string::npos) {
|
||
// During partial parsing, a new tool call may appear with empty name initially
|
||
// The name gets filled in as more input is parsed
|
||
while (msg_accum.tool_calls.size() <= diff.tool_call_index) {
|
||
msg_accum.tool_calls.push_back({ "", "", "" });
|
||
}
|
||
// Always update name and id from diff (may change during incremental parsing), but only if the delta
|
||
// actually contains them
|
||
if (!diff.tool_call_delta.name.empty()) {
|
||
msg_accum.tool_calls[diff.tool_call_index].name = diff.tool_call_delta.name;
|
||
}
|
||
if (!diff.tool_call_delta.id.empty()) {
|
||
msg_accum.tool_calls[diff.tool_call_index].id = diff.tool_call_delta.id;
|
||
}
|
||
if (!diff.tool_call_delta.arguments.empty()) {
|
||
msg_accum.tool_calls[diff.tool_call_index].arguments += diff.tool_call_delta.arguments;
|
||
}
|
||
}
|
||
}
|
||
try {
|
||
assert_msg_equals(msg_current, msg_accum, true);
|
||
} catch (std::exception & e) {
|
||
throw std::runtime_error((std::string("Error comparing accumulated message to current: ") + e.what()).c_str());
|
||
}
|
||
|
||
msg_prev = msg_current;
|
||
}
|
||
|
||
if (!tc.is_partial) {
|
||
assert_msg_equals(tc.expect, parser.parse(tc.input, false), true);
|
||
}
|
||
assert_msg_equals(tc.expect, msg_accum, true);
|
||
}
|
||
|
||
// Global template filter for --template flag
|
||
static std::string g_template_filter;
|
||
|
||
// Fluent builder for PEG parser tests
|
||
class peg_test_builder;
|
||
|
||
class peg_tester {
|
||
common_chat_templates_ptr tmpls_;
|
||
std::string template_path_;
|
||
bool detailed_debug_;
|
||
friend class peg_test_builder;
|
||
|
||
public:
|
||
explicit peg_tester(const std::string & template_path, const bool detailed_debug = false) :
|
||
tmpls_(read_templates(template_path)),
|
||
template_path_(template_path),
|
||
detailed_debug_(detailed_debug) {}
|
||
|
||
const std::string & template_path() const { return template_path_; }
|
||
|
||
peg_test_builder test(const std::string & input);
|
||
};
|
||
|
||
class peg_test_builder {
|
||
peg_tester & tester_;
|
||
peg_test_case tc_;
|
||
|
||
public:
|
||
peg_test_builder(peg_tester & tester, const std::string & input) : tester_(tester) { tc_.input = input; }
|
||
|
||
// Parameter setters
|
||
peg_test_builder & reasoning_format(common_reasoning_format fmt) {
|
||
tc_.params.reasoning_format = fmt;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & tools(std::vector<common_chat_tool> tools) {
|
||
tc_.params.tools = std::move(tools);
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & enable_thinking(bool val) {
|
||
tc_.params.enable_thinking = val;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & parallel_tool_calls(bool val) {
|
||
tc_.params.parallel_tool_calls = val;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & json_schema(const std::string & schema) {
|
||
tc_.params.json_schema = schema;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & is_partial(bool val) {
|
||
tc_.is_partial = val;
|
||
return *this;
|
||
}
|
||
|
||
// Expect setters
|
||
peg_test_builder & expect(const common_chat_msg & msg) {
|
||
tc_.expect = msg;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & expect_content(const std::string & content) {
|
||
tc_.expect.content = content;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & expect_reasoning(const std::string & reasoning) {
|
||
tc_.expect.reasoning_content = reasoning;
|
||
return *this;
|
||
}
|
||
|
||
peg_test_builder & expect_tool_calls(std::vector<common_chat_tool_call> calls) {
|
||
tc_.expect.tool_calls = std::move(calls);
|
||
return *this;
|
||
}
|
||
|
||
// Execute the test
|
||
void run() {
|
||
// Check template filter
|
||
if (!g_template_filter.empty()) {
|
||
// Case-insensitive substring match
|
||
std::string template_path_lower = tester_.template_path();
|
||
std::string filter_lower = g_template_filter;
|
||
std::transform(template_path_lower.begin(), template_path_lower.end(), template_path_lower.begin(),
|
||
::tolower);
|
||
std::transform(filter_lower.begin(), filter_lower.end(), filter_lower.begin(), ::tolower);
|
||
if (template_path_lower.find(filter_lower) == std::string::npos) {
|
||
// Skip this test
|
||
return;
|
||
}
|
||
}
|
||
LOG_INF("\n\x1b[38;5;126m[%s]\x1b[0m\n%s\n\n", tester_.template_path().c_str(), tc_.input.c_str());
|
||
test_peg_parser(tester_.tmpls_.get(), [this](peg_test_case & t) { t = tc_; }, tester_.detailed_debug_);
|
||
}
|
||
};
|
||
|
||
peg_test_builder peg_tester::test(const std::string & input) {
|
||
return peg_test_builder(*this, input);
|
||
}
|
||
|
||
static void test_msgs_oaicompat_json_conversion() {
|
||
LOG_DBG("%s\n", __func__);
|
||
std::vector<common_chat_msg> msgs{
|
||
message_user,
|
||
message_user_parts,
|
||
message_assist_call,
|
||
message_assist_call_thoughts,
|
||
message_assist_call_thoughts_unparsed,
|
||
message_assist_call_thoughts_content,
|
||
message_assist_call_id,
|
||
message_assist_call_idx,
|
||
message_assist_call_python,
|
||
};
|
||
for (const auto & msg : msgs) {
|
||
auto oai_json = common_chat_msgs_to_json_oaicompat({ msg });
|
||
auto msgs2 = common_chat_msgs_parse_oaicompat(oai_json);
|
||
assert_equals((size_t) 1, msgs2.size());
|
||
const auto & msg2 = msgs2[0];
|
||
assert_msg_equals(msg, msg2);
|
||
}
|
||
assert_equals(std::string("[\n"
|
||
" {\n"
|
||
" \"role\": \"user\",\n"
|
||
" \"content\": [\n"
|
||
" {\n"
|
||
" \"type\": \"text\",\n"
|
||
" \"text\": \"Hey\"\n"
|
||
" },\n"
|
||
" {\n"
|
||
" \"type\": \"text\",\n"
|
||
" \"text\": \"there\"\n"
|
||
" }\n"
|
||
" ]\n"
|
||
" }\n"
|
||
"]"),
|
||
common_chat_msgs_to_json_oaicompat({ message_user_parts }).dump(2));
|
||
|
||
// Note: content is "" instead of null due to workaround for templates that render null as "None"
|
||
assert_equals(std::string("[\n"
|
||
" {\n"
|
||
" \"role\": \"assistant\",\n"
|
||
" \"content\": \"\",\n"
|
||
" \"tool_calls\": [\n"
|
||
" {\n"
|
||
" \"type\": \"function\",\n"
|
||
" \"function\": {\n"
|
||
" \"name\": \"python\",\n"
|
||
" \"arguments\": {\n"
|
||
" \"code\": \"print('hey')\"\n"
|
||
" }\n"
|
||
" }\n"
|
||
" }\n"
|
||
" ]\n"
|
||
" }\n"
|
||
"]"),
|
||
common_chat_msgs_to_json_oaicompat({ message_assist_call_python }).dump(2));
|
||
|
||
auto res = common_chat_msgs_parse_oaicompat(json::parse("[{\"role\": \"assistant\", \"tool_calls\": []}]"));
|
||
assert_equals<size_t>(1, res.size());
|
||
assert_equals<std::string>(res[0].role, "assistant");
|
||
assert_equals(true, res[0].content.empty());
|
||
assert_equals(true, res[0].tool_calls.empty());
|
||
|
||
try {
|
||
common_chat_msgs_parse_oaicompat(json::parse("[{\"role\": \"assistant\"}]"));
|
||
throw std::runtime_error("Expected exception");
|
||
} catch (const std::exception & e) {
|
||
if (std::string(e.what()).find("'content'") == std::string::npos) {
|
||
throw std::runtime_error("Expected exception about missing 'content'");
|
||
}
|
||
}
|
||
}
|
||
|
||
static void test_tools_oaicompat_json_conversion() {
|
||
LOG_DBG("%s\n", __func__);
|
||
std::vector<common_chat_tool> tools{
|
||
special_function_tool,
|
||
python_tool,
|
||
};
|
||
|
||
for (const auto & tool : tools) {
|
||
auto oai_json = common_chat_tools_to_json_oaicompat({ tool });
|
||
auto tools2 = common_chat_tools_parse_oaicompat(oai_json);
|
||
assert_equals((size_t) 1, tools2.size());
|
||
auto tool2 = tools2[0];
|
||
assert_equals(tool.name, tool2.name);
|
||
assert_equals(tool.description, tool2.description);
|
||
assert_equals(json::parse(tool.parameters).dump(2), json::parse(tool2.parameters).dump(2));
|
||
}
|
||
|
||
assert_equals(std::string("[\n"
|
||
" {\n"
|
||
" \"type\": \"function\",\n"
|
||
" \"function\": {\n"
|
||
" \"name\": \"special_function\",\n"
|
||
" \"description\": \"I'm special\",\n"
|
||
" \"parameters\": {\n"
|
||
" \"type\": \"object\",\n"
|
||
" \"properties\": {\n"
|
||
" \"arg1\": {\n"
|
||
" \"type\": \"integer\",\n"
|
||
" \"description\": \"The arg.\"\n"
|
||
" }\n"
|
||
" },\n"
|
||
" \"required\": [\n"
|
||
" \"arg1\"\n"
|
||
" ]\n"
|
||
" }\n"
|
||
" }\n"
|
||
" }\n"
|
||
"]"),
|
||
common_chat_tools_to_json_oaicompat({ special_function_tool }).dump(2));
|
||
}
|
||
|
||
static void test_template_output_peg_parsers(bool detailed_debug) {
|
||
LOG_DBG("%s\n", __func__);
|
||
|
||
// JSON schemas
|
||
const char * invoice_schema = R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"amount": {"type": "number"},
|
||
"date": {"type": "string"}
|
||
}
|
||
})";
|
||
|
||
{
|
||
// Ministral-3-14B-Reasoning-2512
|
||
auto tst = peg_tester("models/templates/mistralai-Ministral-3-14B-Reasoning-2512.jinja", detailed_debug);
|
||
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
tst.test("[THINK]I'm\nthinking[/THINK]Hello, world!\nWhat's up?")
|
||
.expect_content("[THINK]I'm\nthinking[/THINK]Hello, world!\nWhat's up?")
|
||
.run();
|
||
|
||
tst.test("[THINK]I'm\nthinking[/THINK]Hello, world!\nWhat's up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
tst.test(R"([TOOL_CALLS]special_function[ARGS]{"arg1":1})")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
tst.test(
|
||
"[THINK]I'm\nthinking[/THINK]"
|
||
R"([TOOL_CALLS]special_function[ARGS]{"arg1":1})")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_thoughts)
|
||
.run();
|
||
|
||
tst.test(R"([TOOL_CALLS]special_function[ARGS]{"arg1": 1})"
|
||
R"([TOOL_CALLS]special_function_with_opt[ARGS]{"arg1": 1, "arg2": 2})")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.parallel_tool_calls(true)
|
||
.tools({
|
||
special_function_tool, special_function_tool_with_optional_param
|
||
})
|
||
.expect_tool_calls({
|
||
{ "special_function", R"({"arg1": 1})", {} },
|
||
{ "special_function_with_opt", R"({"arg1": 1, "arg2": 2})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"[THINK]I need to output the invoice details in JSON[/THINK]"
|
||
"```json\n"
|
||
R"({"amount": 123.45, "date": "2025-12-03"})"
|
||
"\n```")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.json_schema(invoice_schema)
|
||
.expect_reasoning("I need to output the invoice details in JSON")
|
||
.expect_content(R"({"amount": 123.45, "date": "2025-12-03"})")
|
||
.run();
|
||
}
|
||
|
||
{
|
||
// NVIDIA Nemotron-3 Nano
|
||
auto tst = peg_tester("models/templates/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16.jinja", detailed_debug);
|
||
|
||
tst.test("Hello, world!\nWhat's up?").enable_thinking(false).expect(message_assist).run();
|
||
|
||
tst.test("I'm\nthinking\n</think>\nHello, world!\nWhat's up?")
|
||
.enable_thinking(false)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_NONE)
|
||
.expect_content("I'm\nthinking\n</think>\nHello, world!\nWhat's up?")
|
||
.run();
|
||
|
||
tst.test("I'm\nthinking\n</think>\nHello, world!\nWhat's up?")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>\n1\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(false)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
tst.test(
|
||
"I'm\nthinking\n</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>\n1\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_thoughts)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>\n1\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>\n"
|
||
"<tool_call>\n"
|
||
"<function=special_function_with_opt>\n"
|
||
"<parameter=arg1>\n1\n</parameter>\n"
|
||
"<parameter=arg2>\n2\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(false)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.parallel_tool_calls(true)
|
||
.tools({
|
||
special_function_tool, special_function_tool_with_optional_param
|
||
})
|
||
.expect_tool_calls({
|
||
{ "special_function", R"({"arg1": 1})", {} },
|
||
{ "special_function_with_opt", R"({"arg1": 1, "arg2": 2})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=python>\n"
|
||
"<parameter=code>\n"
|
||
"def hello():\n"
|
||
" print(\"Hello, world!\")\n"
|
||
"\n"
|
||
"hello()\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(false)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({
|
||
python_tool
|
||
})
|
||
.expect_tool_calls({
|
||
{ "python", "{\"code\": \"def hello():\\n print(\\\"Hello, world!\\\")\\n\\nhello()\"}", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"I need to output the invoice details in JSON\n"
|
||
"</think>\n"
|
||
R"({"amount": 123.45, "date": "2025-12-03"})")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.enable_thinking(true)
|
||
.json_schema(invoice_schema)
|
||
.expect_reasoning("I need to output the invoice details in JSON")
|
||
.expect_content(R"({"amount": 123.45, "date": "2025-12-03"})")
|
||
.run();
|
||
}
|
||
|
||
{
|
||
// CohereForAI Command-R 7B (2024-tool_use)
|
||
auto tst = peg_tester("models/templates/CohereForAI-c4ai-command-r7b-12-2024-tool_use.jinja", detailed_debug);
|
||
|
||
tst.test("<|START_RESPONSE|>Hello, world!\nWhat's up?<|END_RESPONSE|>").expect(message_assist).run();
|
||
|
||
tst.test(
|
||
"<|START_THINKING|>I'm\nthinking<|END_THINKING|>"
|
||
"<|START_RESPONSE|>Hello, world!\nWhat's up?<|END_RESPONSE|>")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<|START_THINKING|>I'm\nthinking<|END_THINKING|>"
|
||
"<|START_RESPONSE|>Hello, world!\nWhat's up?<|END_RESPONSE|>")
|
||
.expect(message_assist_thoughts_unparsed_r7b)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<|START_THINKING|>I'm\nthinking<|END_THINKING|>"
|
||
"<|START_ACTION|>[\n"
|
||
" {\"tool_call_id\": \"0\", \"tool_name\": \"special_function\", \"parameters\": {\"arg1\": 1}}\n"
|
||
"]<|END_ACTION|>")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_thoughts_call_idx)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<|START_THINKING|>I'm\nthinking<|END_THINKING|>"
|
||
"<|START_ACTION|>[\n"
|
||
" {\"tool_call_id\": \"0\", \"tool_name\": \"special_function\", ")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool })
|
||
.is_partial(true)
|
||
.expect(message_assist_thoughts_partial_call)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<|START_THINKING|><|END_THINKING|>"
|
||
"<|START_ACTION|>[\n"
|
||
" {\"tool_call_id\": \"0\", \"tool_name\": \"special_function\", \"parameters\": {\"arg1\": 1}}\n"
|
||
"]<|END_ACTION|>")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_idx)
|
||
.run();
|
||
}
|
||
|
||
{
|
||
// Google Gemma 2 2B - does not support tool calling
|
||
auto tst = peg_tester("models/templates/google-gemma-2-2b-it.jinja");
|
||
|
||
tst.test("Hello, world!").expect(simple_assist_msg("Hello, world!")).run();
|
||
|
||
tst.test("Line 1\nLine 2\nLine 3").expect(simple_assist_msg("Line 1\nLine 2\nLine 3")).run();
|
||
}
|
||
|
||
{
|
||
// Qwen-QwQ-32B (reasoning model)
|
||
auto tst = peg_tester("models/templates/Qwen-QwQ-32B.jinja");
|
||
|
||
// QwQ always has thinking forced open - input starts after the <think>\n in the prompt
|
||
tst.test("Let me think about this...\n</think>\nThe answer is 42.")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.expect(simple_assist_msg("The answer is 42.", "Let me think about this..."))
|
||
.run();
|
||
|
||
tst.test("Hello, world!").expect(simple_assist_msg("Hello, world!")).run();
|
||
}
|
||
{
|
||
// NousResearch-Hermes-2-Pro and Hermes-3 (tool calling models)
|
||
auto tst = peg_tester("models/templates/NousResearch-Hermes-2-Pro-Llama-3-8B-tool_use.jinja", detailed_debug);
|
||
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}}\n"
|
||
"</tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
tst.test(
|
||
"Hello, world!\nWhat's up?<tool_call>\n"
|
||
"{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}}\n"
|
||
"</tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_content)
|
||
.run();
|
||
|
||
// Note: Hermes template doesn't support thinking/reasoning natively
|
||
// Note: We only support one tool calling format per template, no alternate formats
|
||
}
|
||
{
|
||
// Test simple content-only template
|
||
auto tst = peg_tester("models/templates/google-gemma-2-2b-it.jinja", detailed_debug);
|
||
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
}
|
||
{
|
||
// IBM Granite (reasoning and tool calling model)
|
||
auto tst = peg_tester("models/templates/ibm-granite-granite-3.3-2B-Instruct.jinja", detailed_debug);
|
||
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
tst.test("<think>I'm\nthinking</think>Hello, world!\nWhat's up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
// TODO: pending support for WRAPPED_WITH_REASONING
|
||
// tst.test("<think>I'm\nthinking</think><response>Hello, world!\nWhat's up?</response>")
|
||
// .reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
// .expect(message_assist_thoughts)
|
||
// .run();
|
||
}
|
||
|
||
{
|
||
// ByteDance-Seed-OSS (reasoning and tool calling model)
|
||
auto tst = peg_tester("models/templates/ByteDance-Seed-OSS.jinja", detailed_debug);
|
||
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
tst.test("<seed:think>I'm thinking about the answer</seed:think>Hello, world!")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(simple_assist_msg("Hello, world!", "I'm thinking about the answer"))
|
||
.run();
|
||
|
||
tst.test(
|
||
"<seed:tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>1</parameter>\n"
|
||
"</function>\n"
|
||
"</seed:tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<seed:tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>1</parameter>\n"
|
||
"</function>\n"
|
||
"</seed:tool_call>\n"
|
||
"<seed:tool_call>\n"
|
||
"<function=special_function_with_opt>\n"
|
||
"<parameter=arg1>1</parameter>\n"
|
||
"<parameter=arg2>2</parameter>\n"
|
||
"</function>\n"
|
||
"</seed:tool_call>")
|
||
.parallel_tool_calls(true)
|
||
.tools({
|
||
special_function_tool, special_function_tool_with_optional_param
|
||
})
|
||
.expect_tool_calls({
|
||
{ "special_function", R"({"arg1": 1})", {} },
|
||
{ "special_function_with_opt", R"({"arg1": 1, "arg2": 2})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"<seed:tool_call>\n"
|
||
"<function=todo_list>\n"
|
||
"<parameter=todos>[{\"item\": \"Check stuff\", \"selected\": false}, {\"item\": \"Prepare stuff\", \"selected\": true}]</parameter>\n"
|
||
"</function>\n"
|
||
"</seed:tool_call>")
|
||
.tools({
|
||
todo_list
|
||
})
|
||
.expect_tool_calls({
|
||
{ "todo_list", "{\"todos\": [{\"item\": \"Check stuff\", \"selected\": false}, {\"item\": \"Prepare stuff\", \"selected\": true}]}", {} },
|
||
})
|
||
.run();
|
||
|
||
// single-quote normalization
|
||
tst.test(
|
||
"<seed:tool_call>\n"
|
||
"<function=todo_list>\n"
|
||
"<parameter=todos>[{'item': 'Check stuff', 'selected': false}, {'item': 'Prepare stuff', 'selected': true}]</parameter>\n"
|
||
"</function>\n"
|
||
"</seed:tool_call>")
|
||
.tools({
|
||
todo_list
|
||
})
|
||
.expect_tool_calls({
|
||
{ "todo_list", "{\"todos\": [{\"item\": \"Check stuff\", \"selected\": false}, {\"item\": \"Prepare stuff\", \"selected\": true}]}", {} },
|
||
})
|
||
.run();
|
||
|
||
// tool call with inside quotes
|
||
tst.test(
|
||
"<seed:tool_call>\n"
|
||
"<function=edit>\n"
|
||
"<parameter=filename>\n"
|
||
"foo.cpp\n"
|
||
"</parameter>\n"
|
||
"<parameter=oldString>"
|
||
"def foo(arg = \"14\"):\n"
|
||
" return arg + \"bar\"\n"
|
||
"\n"
|
||
"</parameter>\n"
|
||
"<parameter=newString>"
|
||
"def foo(arg = \"15\"):\n"
|
||
" pass\n"
|
||
"\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</seed:tool_call>")
|
||
.tools({
|
||
edit_tool
|
||
})
|
||
.expect_tool_calls({
|
||
{ "edit", "{\"filename\": \"foo.cpp\", "
|
||
"\"oldString\": \"def foo(arg = \\\"14\\\"):\\n return arg + \\\"bar\\\"\\n\", "
|
||
"\"newString\": \"def foo(arg = \\\"15\\\"):\\n pass\\n\"}", {}
|
||
}
|
||
})
|
||
.run();
|
||
}
|
||
|
||
{
|
||
// Qwen3-Coder (tool calling with XML-style format)
|
||
auto tst = peg_tester("models/templates/Qwen3-Coder.jinja", detailed_debug);
|
||
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>\n"
|
||
"1\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>\n"
|
||
"1\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>\n"
|
||
"<tool_call>\n"
|
||
"<function=special_function_with_opt>\n"
|
||
"<parameter=arg1>\n"
|
||
"1\n"
|
||
"</parameter>\n"
|
||
"<parameter=arg2>\n"
|
||
"2\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.parallel_tool_calls(true)
|
||
.tools({
|
||
special_function_tool, special_function_tool_with_optional_param
|
||
})
|
||
.expect_tool_calls({
|
||
{ "special_function", R"({"arg1": 1})", {} },
|
||
{ "special_function_with_opt", R"({"arg1": 1, "arg2": 2})", {} },
|
||
})
|
||
.run();
|
||
|
||
// Test with code content (multiline)
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=python>\n"
|
||
"<parameter=code>\n"
|
||
"def hello():\n"
|
||
" print(\"Hello, world!\")\n"
|
||
"\n"
|
||
"hello()\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.tools({
|
||
python_tool
|
||
})
|
||
.expect_tool_calls({
|
||
{ "python", "{\"code\": \"def hello():\\n print(\\\"Hello, world!\\\")\\n\\nhello()\"}", {} },
|
||
})
|
||
.run();
|
||
|
||
// Test with HTML tag content
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=html>\n"
|
||
"<parameter=markup>\n"
|
||
"<html>\n"
|
||
" <head>\n"
|
||
" <title>Hello!</title>\n"
|
||
" </head>\n"
|
||
"</html>\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.tools({
|
||
html_tool
|
||
})
|
||
.expect_tool_calls({
|
||
{ "html", "{\"markup\": \"<html>\\n <head>\\n <title>Hello!</title>\\n </head>\\n</html>\"}", {} },
|
||
})
|
||
.run();
|
||
|
||
// Test with TODO list (array of objects)
|
||
tst.test(
|
||
"<tool_call>\n"
|
||
"<function=todo_list>\n"
|
||
"<parameter=todos>\n"
|
||
"[{\"item\": \"Check stuff\", \"selected\": false}, {\"item\": \"Prepare stuff\", \"selected\": true}]\n"
|
||
"</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.tools({
|
||
todo_list
|
||
})
|
||
.expect_tool_calls({
|
||
{ "todo_list", "{\"todos\": [{\"item\": \"Check stuff\", \"selected\": false}, {\"item\": \"Prepare stuff\", \"selected\": true}]}", {} },
|
||
})
|
||
.run();
|
||
}
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-V3.1.jinja", detailed_debug);
|
||
tst.test(
|
||
"<|tool▁calls▁begin|><|tool▁call▁begin|>get_time<|tool▁sep|>{\"city\": "
|
||
"\"XYZCITY\"}<|tool▁call▁end|><|tool▁calls▁end|>")
|
||
.tools({ get_time_tool })
|
||
.expect(message_with_tool_calls("get_time", "{\"city\":\"XYZCITY\"}"))
|
||
.run();
|
||
}
|
||
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-V3.1.jinja", detailed_debug);
|
||
tst.test(
|
||
"REASONING</think><|tool▁calls▁begin|><|tool▁call▁begin|>get_time<|tool▁sep|>{\"city\": "
|
||
"\"Tokyo\"}<|tool▁call▁end|><|tool▁calls▁end|>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ get_time_tool })
|
||
.expect(message_with_tool_calls_and_reasoning("get_time", "{\"city\":\"Tokyo\"}", "REASONING"))
|
||
.run();
|
||
}
|
||
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-V3.1.jinja", detailed_debug);
|
||
tst.test(
|
||
"REASONING</think>CONTENT<|tool▁calls▁begin|><|tool▁call▁begin|>get_time<|tool▁sep|>{\"city\": "
|
||
"\"Paris\"}<|tool▁call▁end|><|tool▁call▁begin|>get_weather<|tool▁sep|>{\"city\": "
|
||
"\"Paris\"}<|tool▁call▁end|><|tool▁calls▁end|>")
|
||
.tools({
|
||
get_time_tool, get_weather_tool
|
||
})
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.parallel_tool_calls(true)
|
||
.expect(message_with_reasoning_content_and_multiple_tool_calls(
|
||
"REASONING", "CONTENT",
|
||
{ { "get_time", "{\"city\":\"Paris\"}" }, { "get_weather", "{\"city\":\"Paris\"}" } }))
|
||
.run();
|
||
}
|
||
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-V3.1.jinja", detailed_debug);
|
||
tst.test("REASONING</think>\nCONTENT")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(simple_assist_msg("CONTENT", "REASONING\n"))
|
||
.run();
|
||
}
|
||
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-V3.1.jinja", detailed_debug);
|
||
tst.test("CONTENT").expect(simple_assist_msg("CONTENT", "")).run();
|
||
}
|
||
|
||
// GLM-4.6 tests - format: <tool_call>function_name\n<arg_key>...</arg_key>\n<arg_value>...</arg_value>\n</tool_call>
|
||
{
|
||
auto tst = peg_tester("models/templates/GLM-4.6.jinja", detailed_debug);
|
||
tst.test(
|
||
"<tool_call>special_function\n"
|
||
"<arg_key>arg1</arg_key>\n<arg_value>1</arg_value>\n"
|
||
"</tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// GLM-4.7-Flash tests - format: <tool_call>function_name<arg_key>...</arg_key><arg_value>...</arg_value></tool_call>
|
||
// Note: Template uses forced-open thinking mode (prompt ends with <think>)
|
||
{
|
||
auto tst = peg_tester("models/templates/GLM-4.7-Flash.jinja", detailed_debug);
|
||
|
||
// Pure content (no reasoning)
|
||
tst.test("Hello, world!\nWhat's up?")
|
||
.enable_thinking(false)
|
||
.expect(message_assist)
|
||
.run();
|
||
|
||
// Reasoning with content (forced-open mode - input starts after <think>)
|
||
tst.test("I'm\nthinking</think>Hello, world!\nWhat's up?")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
// Tool call without reasoning
|
||
tst.test(
|
||
"<tool_call>special_function"
|
||
"<arg_key>arg1</arg_key><arg_value>1</arg_value>"
|
||
"</tool_call>")
|
||
.enable_thinking(false)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
// Tool call with reasoning (forced-open mode)
|
||
tst.test(
|
||
"I'm\nthinking</think>"
|
||
"<tool_call>special_function"
|
||
"<arg_key>arg1</arg_key><arg_value>1</arg_value>"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_thoughts)
|
||
.run();
|
||
|
||
// String argument starting with '[' - should NOT be treated as JSON array
|
||
// This tests the fix for Godot scene files and similar content
|
||
tst.test(
|
||
"<tool_call>html"
|
||
"<arg_key>markup</arg_key><arg_value>[gd_scene load_steps=3 format=3]</arg_value>"
|
||
"</tool_call>")
|
||
.enable_thinking(false)
|
||
.tools({ html_tool })
|
||
.expect_tool_calls({
|
||
{ "html", "{\"markup\": \"[gd_scene load_steps=3 format=3]\"}", {} },
|
||
})
|
||
.run();
|
||
|
||
// Multiple tool calls
|
||
// Note: Parallel tool calls streaming test skipped - the KEY_VALUE_TAGS format has
|
||
// partial parsing edge cases when function names share common prefixes (special_function vs special_function_with_opt)
|
||
// The grammar and full parsing work correctly, but incremental streaming detection needs more work.
|
||
}
|
||
|
||
// Kimi-K2-Thinking tests - FUNC_PREFIXED_INDEXED format
|
||
{
|
||
auto tst = peg_tester("models/templates/Kimi-K2-Thinking.jinja", detailed_debug);
|
||
tst.test(
|
||
"<|tool_calls_section_begin|><|tool_call_begin|>functions.special_function:0<|tool_call_argument_begin|>"
|
||
"{\"arg1\": 1}<|tool_call_end|><|tool_calls_section_end|>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// Apertus-8B-Instruct tests - FUNC_NAME_AS_KEY format
|
||
// Format: <|tools_prefix|>[{"function_name": {...arguments...}}]<|tools_suffix|>
|
||
{
|
||
auto tst = peg_tester("models/templates/Apertus-8B-Instruct.jinja", detailed_debug);
|
||
tst.test("<|tools_prefix|>[{\"special_function\": {\"arg1\": 1}}]<|tools_suffix|>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// MiniMax-M2 tests - XML invoke format with parameter tags
|
||
// Format: <minimax:tool_call><invoke name="func"><parameter name="key">value</parameter></invoke></minimax:tool_call>
|
||
{
|
||
auto tst = peg_tester("models/templates/MiniMax-M2.jinja", detailed_debug);
|
||
tst.test(
|
||
"<minimax:tool_call>\n<invoke name=\"special_function\">\n<parameter "
|
||
"name=\"arg1\">1</parameter>\n</invoke>\n</minimax:tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// NVIDIA-Nemotron-Nano-v2 tests - <TOOLCALL>...</TOOLCALL> format
|
||
// Format: <TOOLCALL>[{"name": "func", "arguments": {...}}]</TOOLCALL>
|
||
{
|
||
auto tst = peg_tester("models/templates/NVIDIA-Nemotron-Nano-v2.jinja", detailed_debug);
|
||
tst.test("<TOOLCALL>[{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}}]</TOOLCALL><SPECIAL_12>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// CohereForAI-c4ai-command-r7b (uses START_RESPONSE/END_RESPONSE, START_THINKING/END_THINKING, START_ACTION/END_ACTION)
|
||
{
|
||
auto tst = peg_tester("models/templates/CohereForAI-c4ai-command-r7b-12-2024-tool_use.jinja", detailed_debug);
|
||
tst.test("<|START_RESPONSE|>Hello, world!\nWhat's up?<|END_RESPONSE|>").expect(message_assist).run();
|
||
tst.test(
|
||
"<|START_THINKING|>I'm\nthinking<|END_THINKING|>"
|
||
"<|START_ACTION|>[\n"
|
||
" {\"tool_call_id\": \"0\", \"tool_name\": \"special_function\", \"parameters\": {\"arg1\": 1}}\n"
|
||
"]<|END_ACTION|>")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_thoughts_call_idx)
|
||
.run();
|
||
}
|
||
// CohereForAI-c4ai-command-r-plus (uses markdown code block format)
|
||
{
|
||
auto tst = peg_tester("models/templates/CohereForAI-c4ai-command-r-plus-tool_use.jinja", detailed_debug);
|
||
tst.test("<|CHATBOT_TOKEN|>Hello, world!\nWhat's up?<|END_OF_TURN_TOKEN|>").expect(message_assist).run();
|
||
// Tool calls: Action: followed by JSON code block
|
||
tst.test(
|
||
"Action:\n"
|
||
"```json\n"
|
||
"[{\"tool_name\": \"special_function\", \"parameters\": {\"arg1\": 1}}]\n"
|
||
"```")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// mistralai-Mistral-Nemo-Instruct-2407.jinja
|
||
{
|
||
auto tst = peg_tester("models/templates/mistralai-Mistral-Nemo-Instruct-2407.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("[TOOL_CALLS][{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}, \"id\": \"123456789\"}]")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_id)
|
||
.run();
|
||
}
|
||
{
|
||
auto tst = peg_tester("models/templates/meetkai-functionary-medium-v3.1.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("<function=special_function>{\"arg1\": 1}</function>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
// Functionary v3.2 - recipient-based format: >>>recipient\n{content}
|
||
{
|
||
auto tst = peg_tester("models/templates/meetkai-functionary-medium-v3.2.jinja", detailed_debug);
|
||
tst.test(">>>all\nHello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test(">>>special_function\n{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// FireFunction
|
||
{
|
||
auto tst = peg_tester("models/templates/fireworks-ai-llama-3-firefunction-v2.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test(" functools[{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}}]")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// DeepSeek R1 Distill Llama 8B - reasoning tests only (forced open thinking)
|
||
// Note: Template uses forced-open mode (prompt ends with <think>), so input shouldn't include opening tag
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-R1-Distill-Llama-8B.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?")
|
||
.enable_thinking(true) // Forced open
|
||
.expect(message_assist)
|
||
.run();
|
||
tst.test("I'm\nthinking</think>Hello, world!\nWhat's up?")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
}
|
||
// llama-cpp DeepSeek R1 template (always forced-open thinking)
|
||
{
|
||
auto tst = peg_tester("models/templates/llama-cpp-deepseek-r1.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("I'm\nthinking</think>Hello, world!\nWhat's up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
tst.test(
|
||
"<|tool▁calls▁begin|><|tool▁call▁begin|>function<|tool▁sep|>special_function\n"
|
||
"```json\n{\"arg1\": 1}```<|tool▁call▁end|><|tool▁calls▁end|>")
|
||
.tools({ special_function_tool })
|
||
.parallel_tool_calls(true)
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
// DeepSeek R1 Distill Qwen 32B - reasoning tests only (forced open thinking)
|
||
// Note: Template uses forced-open mode (prompt ends with <think>), so input shouldn't include opening tag
|
||
{
|
||
auto tst = peg_tester("models/templates/deepseek-ai-DeepSeek-R1-Distill-Qwen-32B.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").enable_thinking(true).expect(message_assist).run();
|
||
tst.test("I'm\nthinking</think>Hello, world!\nWhat's up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
tst.test(
|
||
"<|tool▁calls▁begin|><|tool▁call▁begin|>function<|tool▁sep|>special_function\n"
|
||
"```json\n{\"arg1\": 1}```<|tool▁call▁end|><|tool▁calls▁end|>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
// Kimi-K2 (moonshotai) - FUNC_PREFIXED_INDEXED format
|
||
{
|
||
auto tst = peg_tester("models/templates/moonshotai-Kimi-K2.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test(
|
||
"<|tool_calls_section_begin|><|tool_call_begin|>functions.special_function:0<|tool_call_argument_begin|>"
|
||
"{\"arg1\": 1}<|tool_call_end|><|tool_calls_section_end|>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
// Kimi-K2-Instruct - FUNC_PREFIXED_INDEXED format
|
||
{
|
||
auto tst = peg_tester("models/templates/Kimi-K2-Instruct.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test(
|
||
"<|tool_calls_section_begin|><|tool_call_begin|>functions.special_function:0<|tool_call_argument_begin|>"
|
||
"{\"arg1\": 1}<|tool_call_end|><|tool_calls_section_end|>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// MiMo-VL / Hermes 3 / Qwen 2.5 (Common <tool_call> JSON format)
|
||
for (const auto & path :
|
||
{ "models/templates/MiMo-VL.jinja", "models/templates/NousResearch-Hermes-3-Llama-3.1-8B-tool_use.jinja",
|
||
"models/templates/Qwen-Qwen2.5-7B-Instruct.jinja" }) {
|
||
auto tst = peg_tester(path, detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("<tool_call>\n{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}}\n</tool_call>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// Apriel 1.5
|
||
{
|
||
auto tst = peg_tester("models/templates/unsloth-Apriel-1.5.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("<tool_calls>[{\"name\": \"special_function\", \"arguments\": {\"arg1\": 1}}]</tool_calls>")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
}
|
||
|
||
// Apriel 1.6 Thinker (reasoning-only support)
|
||
{
|
||
auto tst = peg_tester("models/templates/Apriel-1.6-15b-Thinker-fixed.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
// Implicit reasoning start (forced open)
|
||
tst.test("I'm\nthinking\n[BEGIN FINAL RESPONSE]\nHello, world!\nWhat's up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
// Reasoning + Tool calls
|
||
tst.test(
|
||
"I'm\nthinking\n[BEGIN FINAL RESPONSE]\n<tool_calls>[{\"name\": \"special_function\", \"arguments\": "
|
||
"{\"arg1\": 1}}]</tool_calls>")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_thoughts)
|
||
.run();
|
||
}
|
||
|
||
// Mistral Small 3.2 - FUNC_BRACKET_TAG format: [TOOL_CALLS]func_name[CALL_ID]id[ARGS]{...}
|
||
{
|
||
auto tst = peg_tester("models/templates/Mistral-Small-3.2-24B-Instruct-2506.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("[TOOL_CALLS]special_function[CALL_ID]123456789[ARGS]{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_id)
|
||
.run();
|
||
}
|
||
// Devstral - FUNC_BRACKET_TAG format (no ID marker): [TOOL_CALLS]func_name[ARGS]{...}
|
||
{
|
||
auto tst = peg_tester("models/templates/unsloth-mistral-Devstral-Small-2507.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
tst.test("[TOOL_CALLS]special_function[ARGS]{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
tst.test("Hello, world!\nWhat's up?[TOOL_CALLS]special_function[ARGS]{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_content)
|
||
.run();
|
||
}
|
||
|
||
{
|
||
// Llama 3.1
|
||
auto tst = peg_tester("models/templates/meta-llama-Llama-3.1-8B-Instruct.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").tools({ special_function_tool }).expect(message_assist).run();
|
||
}
|
||
|
||
{
|
||
// Llama 3.2
|
||
auto tst = peg_tester("models/templates/meta-llama-Llama-3.2-3B-Instruct.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").tools({ special_function_tool }).expect(message_assist).run();
|
||
}
|
||
|
||
{
|
||
// Llama 3.3
|
||
auto tst = peg_tester("models/templates/meta-llama-Llama-3.3-70B-Instruct.jinja", detailed_debug);
|
||
tst.test("Hello, world!\nWhat's up?").tools({ python_tool }).expect(message_assist).run();
|
||
}
|
||
|
||
// GPT-OSS format tests
|
||
{
|
||
auto tst = peg_tester("models/templates/openai-gpt-oss-120b.jinja", detailed_debug);
|
||
|
||
// Basic content only - final channel
|
||
tst.test("<|channel|>final<|message|>Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
// Basic content only - commentary channel
|
||
tst.test("<|channel|>commentary<|message|>Hello, world!\nWhat's up?").expect(message_assist).run();
|
||
|
||
// Analysis channel (reasoning) with final channel (content)
|
||
tst.test(
|
||
"<|channel|>analysis<|message|>I'm\nthinking<|end|>\n<|channel|>final<|message|>Hello, world!\nWhat's "
|
||
"up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.expect(message_assist_thoughts)
|
||
.run();
|
||
|
||
// Analysis channel only (partial) - still works when reasoning format is set
|
||
tst.test("<|channel|>analysis<|message|>I'm\nthinking")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.is_partial(true)
|
||
.expect_reasoning("I'm\nthinking")
|
||
.run();
|
||
|
||
// Reasoning format none - reasoning stays in content
|
||
tst.test(
|
||
"<|channel|>analysis<|message|>I'm\nthinking<|end|>\n<|channel|>final<|message|>Hello, world!\nWhat's "
|
||
"up?")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_NONE)
|
||
.expect_content(
|
||
"<|channel|>analysis<|message|>I'm\nthinking<|end|>Hello, world!\nWhat's up?")
|
||
.run();
|
||
|
||
// Tool call with recipient in role header: " to=functions.NAME<|channel|>analysis<|message|>JSON"
|
||
tst.test(" to=functions.special_function<|channel|>analysis<|message|>{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
// Tool call with recipient in channel header: "<|channel|>analysis to=functions.NAME<|message|>JSON"
|
||
tst.test("<|channel|>analysis to=functions.special_function<|message|>{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
// Tool call with constraint: " to=functions.NAME<|channel|>analysis <|constrain|>json<|message|>JSON"
|
||
tst.test(" to=functions.special_function<|channel|>analysis <|constrain|>json<|message|>{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
// Tool call in commentary channel (channel header variant)
|
||
tst.test("<|channel|>commentary to=functions.special_function<|message|>{\"arg1\": 1}")
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call)
|
||
.run();
|
||
|
||
// Tool call with reasoning + content (analysis first, then tool call)
|
||
tst.test(
|
||
"<|channel|>analysis<|message|>I'm\nthinking<|end|>\n"
|
||
"<|start|>assistant to=functions.special_function<|channel|>analysis<|message|>{\"arg1\": 1}")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_thoughts)
|
||
.run();
|
||
|
||
// Tool calling with extra channel before
|
||
tst.test(
|
||
"<|channel|>analysis<|message|>I'm\nthinking<|end|><|start|>assistant<|channel|>commentary"
|
||
" to=functions.special_function <|message|>{\"arg1\": 1}")
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({ special_function_tool })
|
||
.expect(message_assist_call_thoughts)
|
||
.run();
|
||
|
||
// Reasoning after final channel
|
||
// Tool calling after final channel
|
||
tst.test(
|
||
"<|channel|>final<|message|><|end|>"
|
||
"<|start|>assistant<|channel|>analysis<|message|>Thinking about edit..."
|
||
)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.expect_reasoning("Thinking about edit...")
|
||
.expect_content("")
|
||
.run();
|
||
|
||
// Tool calling after final channel
|
||
tst.test(
|
||
"<|channel|>final<|message|><|end|>"
|
||
"<|start|>assistant<|channel|>analysis<|message|>Thinking about edit...<|end|>"
|
||
"<|start|>assistant<|channel|>commentary to=functions.edit <|constrain|>json"
|
||
"<|message|>{\"filePath\": \"file.js\", \"oldString\": \"if (part < railCount - 1) {\", \"newString\": \"if (part < 4) {\", \"replaceAll\": false}"
|
||
)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_AUTO)
|
||
.tools({
|
||
{
|
||
/* .name = */ "edit",
|
||
/* .description = */ "Edit a file",
|
||
/* .parameters = */ R"({
|
||
"type": "object",
|
||
"properties": {
|
||
"oldString": {
|
||
"type": "string",
|
||
"description": "Old string to replace."
|
||
},
|
||
"newString": {
|
||
"type": "string",
|
||
"description": "New replacement string."
|
||
},
|
||
"replaceAll": {
|
||
"type": "boolean",
|
||
"description": "Whether to replace all occurences."
|
||
}
|
||
},
|
||
"required": ["oldString", "newString"]
|
||
})",
|
||
}
|
||
})
|
||
.expect_reasoning("Thinking about edit...")
|
||
.expect_tool_calls({
|
||
{ "edit", R"({"filePath": "file.js", "oldString": "if (part < railCount - 1) {", "newString": "if (part < 4) {", "replaceAll": false})", {} }
|
||
})
|
||
.run();
|
||
|
||
// Parallel tool calls
|
||
tst.test(
|
||
" to=functions.special_function<|channel|>analysis<|message|>{\"arg1\": 1}\n"
|
||
"<|start|>assistant to=functions.special_function_with_opt<|channel|>analysis<|message|>{\"arg1\": 1, "
|
||
"\"arg2\": 2}")
|
||
.parallel_tool_calls(true)
|
||
.tools({
|
||
special_function_tool, special_function_tool_with_optional_param
|
||
})
|
||
.expect_tool_calls({
|
||
{ "special_function", R"({"arg1": 1})", {} },
|
||
{ "special_function_with_opt", R"({"arg1": 1, "arg2": 2})", {} },
|
||
})
|
||
.run();
|
||
}
|
||
|
||
{
|
||
auto tst = peg_tester("models/templates/StepFun3.5-Flash.jinja", detailed_debug);
|
||
tst.test("I was thinking</think>Now I'm not.").
|
||
enable_thinking(true).
|
||
reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK).
|
||
expect_reasoning("I was thinking").
|
||
expect_content("Now I'm not.")
|
||
.run();
|
||
|
||
// Test that numeric-looking string values are coerced to strings per the schema
|
||
tst.test(
|
||
"Let me call the magic tool\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=magic>\n"
|
||
"<parameter=name>\nfooBar\n</parameter>\n"
|
||
"<parameter=ref>\n5123123\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ magic_tool })
|
||
.expect_reasoning("Let me call the magic tool")
|
||
.expect_tool_calls({
|
||
{ "magic", R"({"name": "fooBar", "ref": "5123123"})", {} },
|
||
})
|
||
.run();
|
||
|
||
// Test that numeric values are correctly interpreted as numbers when schema calls for number
|
||
tst.test(
|
||
"Let me call the special function\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=special_function>\n"
|
||
"<parameter=arg1>\n42555916\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool })
|
||
.expect_reasoning("Let me call the special function")
|
||
.expect_tool_calls({
|
||
{ "special_function", R"({"arg1": 42555916})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"Let me call the special function with opt\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=special_function_with_opt>\n"
|
||
"<parameter=arg1>\n42555916\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ special_function_tool_with_optional_param })
|
||
.expect_reasoning("Let me call the special function with opt")
|
||
.expect_tool_calls({
|
||
{ "special_function_with_opt", R"({"arg1": 42555916})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"Let me call the magic_int function\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=magic_int>\n"
|
||
"<parameter=ref>\n42555916\n</parameter>\n"
|
||
"<parameter=name>\nbaz\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ magic_int_tool })
|
||
.expect_reasoning("Let me call the magic_int function")
|
||
.expect_tool_calls({
|
||
{ "magic_int", R"({"ref": 42555916, "name": "baz"})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"Call string_param with empty text\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=string_param>\n"
|
||
"<parameter=text>\n\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ string_param_tool })
|
||
.expect_reasoning("Call string_param with empty text")
|
||
.expect_tool_calls({
|
||
{ "string_param", R"({"text": ""})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"Test simple quoted unquoted\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=quoted_unquoted>\n"
|
||
"<parameter=quoted>\n\"foo\"\n</parameter>\n"
|
||
"<parameter=unquoted>\nfoo\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ quoted_unquoted_tool })
|
||
.expect_reasoning("Test simple quoted unquoted")
|
||
.expect_tool_calls({
|
||
{ "quoted_unquoted", R"({"quoted": "\"foo\"", "unquoted": "foo"})", {} },
|
||
})
|
||
.run();
|
||
|
||
tst.test(
|
||
"Test complex quoted unquoted\n"
|
||
"</think>\n"
|
||
"<tool_call>\n"
|
||
"<function=quoted_unquoted>\n"
|
||
"<parameter=quoted>\n\"printf(\\\"foo\\\");\"\n</parameter>\n"
|
||
"<parameter=unquoted>\nprintf(\"foo\");\n</parameter>\n"
|
||
"</function>\n"
|
||
"</tool_call>")
|
||
.enable_thinking(true)
|
||
.reasoning_format(COMMON_REASONING_FORMAT_DEEPSEEK)
|
||
.tools({ quoted_unquoted_tool })
|
||
.expect_reasoning("Test complex quoted unquoted")
|
||
.expect_tool_calls({
|
||
{ "quoted_unquoted", R"({ "quoted" : "\"printf(\\\"foo\\\");\"", "unquoted": "printf(\"foo\");" })", {} }
|
||
})
|
||
.run();
|
||
|
||
}
|
||
}
|
||
|
||
static void test_msg_diffs_compute() {
|
||
LOG_DBG("%s\n", __func__);
|
||
{
|
||
common_chat_msg msg1;
|
||
|
||
common_chat_msg msg2;
|
||
msg2.content = "Hello, world!";
|
||
|
||
common_chat_msg_diff diff;
|
||
diff.content_delta = "Hello, world!";
|
||
|
||
assert_equals({ diff }, common_chat_msg_diff::compute_diffs(msg1, msg2));
|
||
}
|
||
{
|
||
common_chat_msg msg1;
|
||
msg1.content = "Hello,";
|
||
|
||
common_chat_msg msg2;
|
||
msg2.content = "Hello, world!";
|
||
|
||
common_chat_msg_diff diff;
|
||
diff.content_delta = " world!";
|
||
|
||
assert_equals({ diff }, common_chat_msg_diff::compute_diffs(msg1, msg2));
|
||
}
|
||
{
|
||
common_chat_msg msg0;
|
||
|
||
common_chat_msg msg1;
|
||
msg1.tool_calls = {
|
||
{ "special_function", "{\"ar", /* .id = */ "123" }
|
||
};
|
||
|
||
common_chat_msg msg2;
|
||
msg2.tool_calls = {
|
||
{ "special_function", "{\"arg1\": 1}", /* .id = */ "123" }
|
||
};
|
||
|
||
common_chat_msg_diff diff01;
|
||
diff01.tool_call_index = 0;
|
||
diff01.tool_call_delta.name = "special_function";
|
||
diff01.tool_call_delta.id = "123";
|
||
diff01.tool_call_delta.arguments = "{\"ar";
|
||
|
||
assert_equals({ diff01 }, common_chat_msg_diff::compute_diffs(msg0, msg1));
|
||
|
||
common_chat_msg_diff diff12;
|
||
diff12.tool_call_index = 0;
|
||
// Note: neither id nor name change here.
|
||
diff12.tool_call_delta.arguments = "g1\": 1}";
|
||
|
||
assert_equals({ diff12 }, common_chat_msg_diff::compute_diffs(msg1, msg2));
|
||
}
|
||
{
|
||
common_chat_msg msg0;
|
||
|
||
common_chat_msg msg2;
|
||
msg2.tool_calls = {
|
||
{ "f1", "{\"arg1\": 1}", /* .id = */ "123" },
|
||
{ "f2", "{\"arg2\": 2}", /* .id = */ "222" },
|
||
};
|
||
|
||
common_chat_msg_diff diff1;
|
||
diff1.tool_call_index = 0;
|
||
diff1.tool_call_delta.name = "f1";
|
||
diff1.tool_call_delta.id = "123";
|
||
diff1.tool_call_delta.arguments = "{\"arg1\": 1}";
|
||
|
||
common_chat_msg_diff diff2;
|
||
diff2.tool_call_index = 1;
|
||
diff2.tool_call_delta.name = "f2";
|
||
diff2.tool_call_delta.id = "222";
|
||
diff2.tool_call_delta.arguments = "{\"arg2\": 2}";
|
||
|
||
assert_equals({ diff1, diff2 }, common_chat_msg_diff::compute_diffs(msg0, msg2));
|
||
}
|
||
}
|
||
|
||
int main(int argc, char ** argv) {
|
||
bool detailed_debug = false;
|
||
bool only_run_filtered = false;
|
||
|
||
// Check for --template flag
|
||
for (int i = 1; i < argc; i++) {
|
||
std::string arg = argv[i];
|
||
if (arg == "--template" && i + 1 < argc) {
|
||
g_template_filter = argv[++i];
|
||
// Only run PEG parser tests with the filter
|
||
only_run_filtered = true;
|
||
}
|
||
if (arg == "--detailed") {
|
||
detailed_debug = true;
|
||
common_log_set_verbosity_thold(999);
|
||
}
|
||
}
|
||
|
||
if (only_run_filtered) {
|
||
test_template_output_peg_parsers(detailed_debug);
|
||
std::cout << "\n[chat] All template tests passed!" << '\n';
|
||
return 0;
|
||
}
|
||
|
||
#ifndef _WIN32
|
||
if (argc > 1) {
|
||
common_chat_templates_inputs inputs;
|
||
common_chat_msg msg;
|
||
msg.role = "user";
|
||
msg.content = "Hey";
|
||
inputs.messages = { msg };
|
||
inputs.tools = { special_function_tool };
|
||
|
||
std::cout << "| Template | Format |\n";
|
||
std::cout << "|----------|--------|\n";
|
||
|
||
for (int i = 1; i < argc; i++) {
|
||
try {
|
||
std::string path = argv[i];
|
||
if (path.rfind(".jinja") != path.size() - 6) {
|
||
std::cerr << "Skipping non-jinja file: " << path << '\n';
|
||
continue;
|
||
}
|
||
auto tmpls = read_templates(path);
|
||
auto parts = string_split(path, "/");
|
||
const auto & name = parts[parts.size() - 1];
|
||
const auto * format = common_chat_format_name(common_chat_templates_apply(tmpls.get(), inputs).format);
|
||
std::cout << "| " << name << " | " << format << " |\n";
|
||
} catch (const std::exception & e) {
|
||
std::cerr << "Failed to process " << argv[i] << ": " << e.what() << '\n';
|
||
}
|
||
}
|
||
} else
|
||
#endif
|
||
{
|
||
test_msg_diffs_compute();
|
||
test_msgs_oaicompat_json_conversion();
|
||
test_tools_oaicompat_json_conversion();
|
||
test_template_output_peg_parsers(detailed_debug);
|
||
std::cout << "\n[chat] All tests passed!" << '\n';
|
||
}
|
||
return 0;
|
||
}
|