Revert "sampling : stop short if backend sampler sampled a token"
This reverts commit 87b2719eca.
This commit is contained in:
parent
c0b182f4d6
commit
10bd640aae
|
|
@ -2100,7 +2100,6 @@ void llm_graph_context::build_sampling() const {
|
|||
if (data.sampled != nullptr) {
|
||||
res->t_sampled[seq_id] = data.sampled;
|
||||
ggml_build_forward_expand(gf, data.sampled);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (data.probs != nullptr) {
|
||||
|
|
|
|||
|
|
@ -441,8 +441,6 @@ static void test_backend_temp_sampling(const char * model_path) {
|
|||
struct llama_sampler_chain_params backend_chain_params = llama_sampler_chain_default_params();
|
||||
struct llama_sampler * backend_sampler_chain = llama_sampler_chain_init(backend_chain_params);
|
||||
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_temp(temp));
|
||||
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_top_k(40));
|
||||
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_dist(18));
|
||||
|
||||
std::vector<llama_sampler_seq_config> backend_sampler_configs = {
|
||||
{ seq_id, backend_sampler_chain },
|
||||
|
|
|
|||
Loading…
Reference in New Issue