Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions test/test_cli_mode.py
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,7 @@ def test_cli_mode_uses_llamacpp_when_available(self):

assert result == "CLI response from DeepSeek V3"
self.mock_available.assert_called_once_with(TEST_LLAMACPP_MODEL)
self.mock_chat_llamacpp.assert_called_once_with(TEST_LLAMACPP_MODEL, 'Write a function')
self.mock_chat_llamacpp.assert_called_once_with(TEST_LLAMACPP_MODEL, 'Write a function', None)

def test_cli_mode_fallback_to_ollama_when_unavailable(self):
"""Test CLI mode falls back to ollama when model not available in llama.cpp."""
Expand All @@ -119,7 +119,7 @@ def test_cli_mode_fallback_to_ollama_when_unavailable(self):

assert result == "Ollama response from DeepSeek Coder"
self.mock_available.assert_called_once_with(TEST_OLLAMA_MODEL)
self.mock_chat_ollama.assert_called_once_with(TEST_OLLAMA_MODEL, 'Help with coding')
self.mock_chat_ollama.assert_called_once_with(TEST_OLLAMA_MODEL, 'Help with coding', None)

def test_default_mode_is_cli(self):
"""Test that default mode is CLI when no llama_mode specified."""
Expand All @@ -130,7 +130,7 @@ def test_default_mode_is_cli(self):

assert result == "Default CLI mode response"
self.mock_available.assert_called_once_with(TEST_LLAMACPP_MODEL)
self.mock_chat_llamacpp.assert_called_once_with(TEST_LLAMACPP_MODEL, 'Help me')
self.mock_chat_llamacpp.assert_called_once_with(TEST_LLAMACPP_MODEL, 'Help me', None)


class TestCLIModeIntegration:
Expand Down
4 changes: 2 additions & 2 deletions test/test_server_mode.py
Original file line number Diff line number Diff line change
Expand Up @@ -111,7 +111,7 @@ def test_server_mode_uses_llamacpp_when_available(self):

assert result == "Server response from DeepSeek V3"
self.mock_available.assert_called_once_with(TEST_LLAMACPP_MODEL)
self.mock_chat_server.assert_called_once_with(TEST_LLAMACPP_MODEL, 'Explain code')
self.mock_chat_server.assert_called_once_with(TEST_LLAMACPP_MODEL, 'Explain code', None)

def test_server_mode_fallback_to_ollama_when_unavailable(self):
"""Test server mode falls back to ollama when model not available in llama.cpp."""
Expand All @@ -122,7 +122,7 @@ def test_server_mode_fallback_to_ollama_when_unavailable(self):

assert result == "Ollama fallback response"
self.mock_available.assert_called_once_with(TEST_OLLAMA_MODEL)
self.mock_chat_ollama.assert_called_once_with(TEST_OLLAMA_MODEL, 'Debug code')
self.mock_chat_ollama.assert_called_once_with(TEST_OLLAMA_MODEL, 'Debug code', None)

def test_server_mode_requires_server_url(self):
"""Test server mode requires LLAMA_SERVER_URL to be set."""
Expand Down