@@ -249,27 +249,29 @@ static void test_llama_kv_cache_hybrid_constructor() {
249249
250250 std::unique_ptr<llama_kv_cache_unified> u_cache (
251251 new llama_kv_cache_unified (
252- /* model */ *model,
253- /* filter */ unified_filter,
254- /* type_k */ GGML_TYPE_F32,
255- /* type_v */ GGML_TYPE_F16,
256- /* v_trans */ false ,
257- /* offload */ false ,
258- /* kv_size */ 10 ,
259- /* padding */ 10 ,
260- /* n_swa */ 0 ,
261- /* swa_type */ LLAMA_SWA_TYPE_NONE
252+ /* model */ *model,
253+ /* filter */ unified_filter,
254+ /* type_k */ GGML_TYPE_F32,
255+ /* type_v */ GGML_TYPE_F16,
256+ /* v_trans */ false ,
257+ /* offload */ false ,
258+ /* kv_size */ 10 ,
259+ /* n_seq_max */ 1 ,
260+ /* padding */ 10 ,
261+ /* n_swa */ 0 ,
262+ /* swa_type */ LLAMA_SWA_TYPE_NONE
262263 )
263264 );
264265 auto * u_cache_ptr = u_cache.get ();
265266 std::unique_ptr<llama_kv_cache_recurrent> r_cache (
266267 new llama_kv_cache_recurrent (
267- /* model */ *model,
268- /* filter */ recurrent_filter,
269- /* type_k */ GGML_TYPE_F32,
270- /* type_v */ GGML_TYPE_F16,
271- /* offload */ false ,
272- /* kv_size */ 10
268+ /* model */ *model,
269+ /* filter */ recurrent_filter,
270+ /* type_k */ GGML_TYPE_F32,
271+ /* type_v */ GGML_TYPE_F16,
272+ /* offload */ false ,
273+ /* kv_size */ 10 ,
274+ /* n_seq_max */ 1
273275 )
274276 );
275277 auto * r_cache_ptr = r_cache.get ();
0 commit comments