- ggml_type_traits_t (80 -> 72 bytes) - llama_batch (72 -> 64 bytes) - llama_model_params (56 -> 48 bytes) - hash_node (32 -> 24 bytes) - ggml_compute_state (32 -> 24 bytes) - gguf_tensor_info (88 -> 80 bytes) |
||
|---|---|---|
| .. | ||
| CMakeLists.txt | ||
| README.md | ||
| parallel.cpp | ||
README.md
llama.cpp/example/parallel
Simplified simulation of serving incoming requests in parallel