Revert "sampling : stop short if backend sampler sampled a token"
This reverts commit 87b2719eca.
This commit is contained in:
parent
c0b182f4d6
commit
10bd640aae
|
|
@ -2100,7 +2100,6 @@ void llm_graph_context::build_sampling() const {
|
||||||
if (data.sampled != nullptr) {
|
if (data.sampled != nullptr) {
|
||||||
res->t_sampled[seq_id] = data.sampled;
|
res->t_sampled[seq_id] = data.sampled;
|
||||||
ggml_build_forward_expand(gf, data.sampled);
|
ggml_build_forward_expand(gf, data.sampled);
|
||||||
continue;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (data.probs != nullptr) {
|
if (data.probs != nullptr) {
|
||||||
|
|
|
||||||
|
|
@ -441,8 +441,6 @@ static void test_backend_temp_sampling(const char * model_path) {
|
||||||
struct llama_sampler_chain_params backend_chain_params = llama_sampler_chain_default_params();
|
struct llama_sampler_chain_params backend_chain_params = llama_sampler_chain_default_params();
|
||||||
struct llama_sampler * backend_sampler_chain = llama_sampler_chain_init(backend_chain_params);
|
struct llama_sampler * backend_sampler_chain = llama_sampler_chain_init(backend_chain_params);
|
||||||
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_temp(temp));
|
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_temp(temp));
|
||||||
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_top_k(40));
|
|
||||||
llama_sampler_chain_add(backend_sampler_chain, llama_sampler_init_dist(18));
|
|
||||||
|
|
||||||
std::vector<llama_sampler_seq_config> backend_sampler_configs = {
|
std::vector<llama_sampler_seq_config> backend_sampler_configs = {
|
||||||
{ seq_id, backend_sampler_chain },
|
{ seq_id, backend_sampler_chain },
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue