llama : bump max seq limit from 64 to 256 (#15916)

ggml-ci
This commit is contained in:
Georgi Gerganov 2025-09-18 12:47:56 +03:00 committed by GitHub
parent b213fce89b
commit e58174cecb
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
1 changed files with 1 additions and 1 deletions

View File

@ -4,7 +4,7 @@
#include <cstdint> #include <cstdint>
#define LLAMA_MAX_SEQ 64 #define LLAMA_MAX_SEQ 256
struct llama_cparams { struct llama_cparams {
uint32_t n_ctx; // context size used during inference uint32_t n_ctx; // context size used during inference