@@ -26,14 +26,15 @@ def test_token_counter(self):
2626 ]
2727
2828 # check if test covers all models (increase if new models are added)
29- assert len (OpenAIChatModel .__args__ ) == 16 # type: ignore
29+ assert len (OpenAIChatModel .__args__ ) == 17 # type: ignore
3030
3131 client = AsyncTypeOpenAI (api_key = "mock" )
3232
3333 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo" ) == 27
3434 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo-0301" ) == 29
3535 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo-0613" ) == 27
3636 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo-1106" ) == 27
37+ assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo-0125" ) == 27
3738 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo-16k" ) == 27
3839 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-3.5-turbo-16k-0613" ) == 27
3940 assert client .chat .completions .num_tokens_from_messages (test_messages , model = "gpt-4" ) == 27
@@ -49,14 +50,15 @@ def test_token_counter(self):
4950
5051 def test_max_token_counter (self ):
5152 # check if test covers all models (increase if new models are added)
52- assert len (OpenAIChatModel .__args__ ) == 16 # type: ignore
53+ assert len (OpenAIChatModel .__args__ ) == 17 # type: ignore
5354
5455 client = AsyncTypeOpenAI (api_key = "mock" )
5556
5657 assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo" ) == 4096
5758 assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo-0301" ) == 4096
5859 assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo-0613" ) == 4096
5960 assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo-1106" ) == 16384
61+ assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo-0125" ) == 16384
6062 assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo-16k" ) == 16384
6163 assert client .chat .completions .max_tokens_of_model ("gpt-3.5-turbo-16k-0613" ) == 16384
6264 assert client .chat .completions .max_tokens_of_model ("gpt-4" ) == 8192
0 commit comments