We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent adea1c9 commit cfebb6eCopy full SHA for cfebb6e
src/llama-graph.cpp
@@ -1243,8 +1243,9 @@ llm_graph_input_attn_no_cache * llm_graph_context::build_attn_inp_no_cache() con
1243
}
1244
1245
llm_graph_input_attn_no_cache * llm_graph_context::build_attn_inp_no_cache_iswa() const {
1246
- // Default sliding window size - can be made configurable via cparams
1247
- const int n_swa = 128;
+ // Use the sliding window size from hyperparameters
+ // If hparams.n_swa is 0, use a default value (128)
1248
+ const int n_swa = hparams.n_swa > 0 ? hparams.n_swa : 128;
1249
1250
auto inp = std::make_unique<llm_graph_input_attn_no_cache>(hparams, cparams, n_swa);
1251
0 commit comments