Skip to content

Commit 4bf6181

Browse files
committed
add comment about not reserving graphs with zero outputs
1 parent af61b63 commit 4bf6181

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

src/llama-context.cpp

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -285,6 +285,7 @@ llama_context::llama_context(
285285
const uint32_t n_seqs = cparams.kv_unified ? 1 : cparams.n_seq_max;
286286
const uint32_t n_tokens = std::min(cparams.n_ctx, cparams.n_ubatch);
287287

288+
// avoid reserving graphs with zero outputs
288289
n_outputs = 1;
289290

290291
LLAMA_LOG_DEBUG("%s: worst-case: n_tokens = %d, n_seqs = %d, n_outputs = %d\n", __func__, n_tokens, n_seqs, n_outputs);

0 commit comments

Comments
 (0)