Skip to content

Commit 940af8d

Browse files
committed
cont : future-proof the swa pad
1 parent d2c30c6 commit 940af8d

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

src/llama-kv-cache-iswa.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@ llama_kv_cache_iswa::llama_kv_cache_iswa(
4747

4848
// note: the SWA cache is always padded to 256 for performance
4949
// https://github.com/ggml-org/llama.cpp/issues/17037
50-
uint32_t size_swa = std::min(size_base, GGML_PAD(hparams.n_swa*(unified ? n_seq_max : 1) + n_ubatch, 256));
50+
uint32_t size_swa = GGML_PAD(std::min(size_base, hparams.n_swa*(unified ? n_seq_max : 1) + n_ubatch), 256);
5151

5252
// when using full-size SWA cache, we set the SWA cache size to be equal to the base cache size
5353
if (swa_full) {

0 commit comments

Comments
 (0)