@@ -108,7 +108,7 @@ def test_cli_mode_uses_llamacpp_when_available(self):
108108
109109 assert result == "CLI response from DeepSeek V3"
110110 self .mock_available .assert_called_once_with (TEST_LLAMACPP_MODEL )
111- self .mock_chat_llamacpp .assert_called_once_with (TEST_LLAMACPP_MODEL , 'Write a function' , None , None )
111+ self .mock_chat_llamacpp .assert_called_once_with (TEST_LLAMACPP_MODEL , 'Write a function' , system_prompt = None , image_files = None )
112112
113113 def test_cli_mode_fallback_to_ollama_when_unavailable (self ):
114114 """Test CLI mode falls back to ollama when model not available in llama.cpp."""
@@ -119,7 +119,7 @@ def test_cli_mode_fallback_to_ollama_when_unavailable(self):
119119
120120 assert result == "Ollama response from DeepSeek Coder"
121121 self .mock_available .assert_called_once_with (TEST_OLLAMA_MODEL )
122- self .mock_chat_ollama .assert_called_once_with (TEST_OLLAMA_MODEL , 'Help with coding' , None , None )
122+ self .mock_chat_ollama .assert_called_once_with (TEST_OLLAMA_MODEL , 'Help with coding' , system_prompt = None , image_files = None )
123123
124124 def test_default_mode_is_cli (self ):
125125 """Test that default mode is CLI when no llama_mode specified."""
@@ -130,7 +130,7 @@ def test_default_mode_is_cli(self):
130130
131131 assert result == "Default CLI mode response"
132132 self .mock_available .assert_called_once_with (TEST_LLAMACPP_MODEL )
133- self .mock_chat_llamacpp .assert_called_once_with (TEST_LLAMACPP_MODEL , 'Help me' , None , None )
133+ self .mock_chat_llamacpp .assert_called_once_with (TEST_LLAMACPP_MODEL , 'Help me' , system_prompt = None , image_files = None )
134134
135135
136136class TestCLIModeIntegration :
0 commit comments