227 lines
6.2 KiB
C++
227 lines
6.2 KiB
C++
#include "arg.h"
|
|
#include "common.h"
|
|
#include "log.h"
|
|
#include "llama-cpp.h"
|
|
#include "../src/llama-ext.h"
|
|
#include "ggml.h"
|
|
#include "gguf-model-data.h"
|
|
#include "gguf.h"
|
|
#include "ggml-backend.h"
|
|
#include "download.h"
|
|
|
|
#include <array>
|
|
#include <vector>
|
|
#include <set>
|
|
#include <fstream>
|
|
#include <iostream>
|
|
#include <random>
|
|
|
|
// Noop because weights are not needed
|
|
static void set_tensor_data(struct ggml_tensor * tensor, void * userdata) {
|
|
GGML_UNUSED(tensor);
|
|
GGML_UNUSED(userdata);
|
|
}
|
|
|
|
struct input_tensor {
|
|
ggml_type type;
|
|
std::array<int64_t, 4> ne;
|
|
std::array<size_t, 4> nb;
|
|
|
|
input_tensor(ggml_type type, int64_t * ne, size_t * nb): type(type) {
|
|
memcpy(this->ne.data(), ne, 4 * sizeof(int64_t));
|
|
memcpy(this->nb.data(), nb, 4 * sizeof(size_t));
|
|
}
|
|
|
|
bool operator<(const input_tensor &b) const {
|
|
return std::tie(type, ne, nb) <
|
|
std::tie(b.type, b.ne, b.nb);
|
|
}
|
|
|
|
void serialize(std::ostream& out) const {
|
|
out << type << ' ';
|
|
for (size_t i = 0; i < 4; i++) {
|
|
out << ne[i] << ' ';
|
|
}
|
|
for (size_t i = 0; i < 4; i++) {
|
|
out << nb[i] << ' ';
|
|
}
|
|
}
|
|
};
|
|
|
|
struct test_object {
|
|
ggml_op op;
|
|
ggml_type type;
|
|
std::array<int64_t, 4> ne;
|
|
std::vector<int32_t> op_params;
|
|
std::vector<input_tensor> sources;
|
|
std::string name;
|
|
|
|
void serialize(std::ostream& out) const {
|
|
out << op << ' ' << type << ' ';
|
|
for (size_t i = 0; i < 4; i++) {
|
|
out << ne[i] << ' ';
|
|
}
|
|
|
|
out << op_params.size() << ' ';
|
|
for (size_t i = 0; i < op_params.size(); i++) {
|
|
out << op_params[i] << ' ';
|
|
}
|
|
|
|
out << sources.size() << ' ';
|
|
for (size_t s = 0; s < sources.size(); s++) {
|
|
sources[s].serialize(out);
|
|
}
|
|
|
|
if (!name.empty()) {
|
|
out << name;
|
|
} else {
|
|
out << '-';
|
|
}
|
|
|
|
out << '\n';
|
|
}
|
|
|
|
bool operator<(const test_object &b) const {
|
|
return std::tie(op, type, ne, op_params, sources) <
|
|
std::tie(b.op, b.type, b.ne, b.op_params, b.sources);
|
|
}
|
|
};
|
|
|
|
static void extract_graph_ops(ggml_cgraph * cgraph, const char * label, std::set<test_object> & tests) {
|
|
int n_nodes = ggml_graph_n_nodes(cgraph);
|
|
int n_skipped = 0;
|
|
int n_before = (int) tests.size();
|
|
for (int i = 0; i < n_nodes; i++) {
|
|
ggml_tensor * node = ggml_graph_node(cgraph, i);
|
|
|
|
if (node->op == GGML_OP_NONE || node->op == GGML_OP_VIEW || node->op == GGML_OP_RESHAPE || node->op == GGML_OP_PERMUTE || node->op == GGML_OP_TRANSPOSE) {
|
|
n_skipped++;
|
|
continue;
|
|
}
|
|
|
|
test_object test;
|
|
|
|
test.op = node->op;
|
|
test.type = node->type;
|
|
memcpy(&test.ne, node->ne, 4 * sizeof(int64_t));
|
|
|
|
test.op_params.resize(GGML_MAX_OP_PARAMS / sizeof(int32_t));
|
|
memcpy(test.op_params.data(), node->op_params, GGML_MAX_OP_PARAMS);
|
|
|
|
for (size_t s = 0; s < GGML_MAX_SRC; s++) {
|
|
if (node->src[s] == nullptr) {
|
|
break;
|
|
}
|
|
|
|
test.sources.emplace_back(node->src[s]->type, node->src[s]->ne, node->src[s]->nb);
|
|
}
|
|
|
|
test.name = node->name;
|
|
tests.insert(test);
|
|
}
|
|
|
|
int n_new = (int) tests.size() - n_before;
|
|
LOG_INF("%s: %d unique ops, %d total nodes, %d skipped (view ops)\n",
|
|
label, n_new, n_nodes, n_skipped);
|
|
}
|
|
|
|
int main(int argc, char ** argv) {
|
|
common_params params;
|
|
params.out_file = "tests.txt";
|
|
|
|
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_EXPORT_GRAPH_OPS)) {
|
|
return 1;
|
|
}
|
|
|
|
common_init();
|
|
|
|
// Load CPU-only
|
|
ggml_backend_dev_t cpu_device = ggml_backend_dev_by_type(GGML_BACKEND_DEVICE_TYPE_CPU);
|
|
params.devices = { cpu_device, nullptr };
|
|
params.fit_params = false;
|
|
params.n_gpu_layers = 0;
|
|
|
|
params.warmup = false;
|
|
|
|
llama_context * ctx;
|
|
common_init_result_ptr init_result;
|
|
llama_context_ptr ctx2;
|
|
llama_model_ptr model;
|
|
|
|
if (params.model.hf_repo.empty()) {
|
|
init_result = common_init_from_params(params);
|
|
|
|
ctx = init_result->context();
|
|
} else {
|
|
#ifdef LLAMA_HF_FETCH
|
|
auto [hf_repo, hf_quant] = common_download_split_repo_tag(params.model.hf_repo);
|
|
if (hf_quant.empty() || hf_quant == "latest") {
|
|
hf_quant = "Q4_K_M";
|
|
}
|
|
|
|
gguf_context_ptr gguf_ctx = gguf_fetch_gguf_ctx(hf_repo, hf_quant);
|
|
if (!gguf_ctx) {
|
|
LOG_ERR("failed to fetch GGUF metadata from %s\n", hf_repo.c_str());
|
|
return 1;
|
|
}
|
|
|
|
llama_model_params model_params = llama_model_default_params();
|
|
model_params.devices = params.devices.data();
|
|
model_params.no_alloc = true;
|
|
|
|
model.reset(llama_model_init_from_user(gguf_ctx.get(), set_tensor_data, nullptr, model_params));
|
|
|
|
if (!model) {
|
|
LOG_ERR("failed to create llama_model from %s\n", hf_repo.c_str());
|
|
return 1;
|
|
}
|
|
|
|
llama_context_params ctx_params = llama_context_default_params();
|
|
ctx2.reset(llama_init_from_model(model.get(), ctx_params));
|
|
ctx = ctx2.get();
|
|
|
|
if (!ctx) {
|
|
LOG_ERR("failed to create llama_context\n");
|
|
return 1;
|
|
}
|
|
#else
|
|
LOG_ERR("export-graph-ops compiled without HF fetch support\n");
|
|
return 1;
|
|
#endif
|
|
}
|
|
|
|
const uint32_t n_seqs = llama_n_seq_max(ctx);
|
|
const uint32_t n_tokens = std::min(llama_n_ctx(ctx), llama_n_ubatch(ctx));
|
|
|
|
std::set<test_object> tests;
|
|
|
|
auto * gf_pp = llama_graph_reserve(ctx, n_tokens, n_seqs, n_tokens);
|
|
if (!gf_pp) {
|
|
LOG_ERR("failed to reserve prompt processing graph\n");
|
|
return 1;
|
|
}
|
|
extract_graph_ops(gf_pp, "pp", tests);
|
|
|
|
auto * gf_tg = llama_graph_reserve(ctx, n_seqs, n_seqs, n_seqs);
|
|
if (!gf_tg) {
|
|
LOG_ERR("failed to reserve token generation graph\n");
|
|
return 1;
|
|
}
|
|
extract_graph_ops(gf_tg, "tg", tests);
|
|
|
|
LOG_INF("%d unique ops total\n", (int) tests.size());
|
|
|
|
std::ofstream f(params.out_file);
|
|
|
|
if (!f.is_open()) {
|
|
LOG_ERR("unable to open output file: %s\n", params.out_file.c_str());
|
|
return 1;
|
|
}
|
|
|
|
for (const auto& test : tests) {
|
|
test.serialize(f);
|
|
}
|
|
|
|
return 0;
|
|
}
|