Skip to content

Commit b135927

Browse files
committed
llama : fix missing worst case flag during reserve
1 parent dec6ce2 commit b135927

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

src/llama.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19694,7 +19694,7 @@ struct llama_context * llama_new_context_with_model(
1969419694
int n_nodes_tg = ggml_graph_n_nodes(gf_tg);
1969519695

1969619696
// reserve again with pp graph to avoid ggml-alloc reallocations during inference
19697-
gf_pp = llama_build_graph(*ctx, ubatch_pp, false);
19697+
gf_pp = llama_build_graph(*ctx, ubatch_pp, true);
1969819698
if (!ggml_backend_sched_reserve(ctx->sched, gf_pp)) {
1969919699
LLAMA_LOG_ERROR("%s: failed to allocate compute buffers\n", __func__);
1970019700
llama_free(ctx);

0 commit comments

Comments
 (0)