Skip to content

Commit 7656cb3

Browse files
committed
test fix
1 parent 2331fb4 commit 7656cb3

8 files changed

+12
-12
lines changed

tests/local_testing/test_function_calling.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ def get_current_weather(location, unit="fahrenheit"):
5050
"claude-3-haiku-20240307",
5151
"gemini/gemini-1.5-pro",
5252
"anthropic.claude-3-sonnet-20240229-v1:0",
53-
"groq/llama3-8b-8192",
53+
"groq/llama-3.1-8b-instant",
5454
"cohere_chat/command-r",
5555
],
5656
)

tests/local_testing/test_lowest_cost_routing.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ async def test_get_available_deployments():
3131
},
3232
{
3333
"model_name": "gpt-3.5-turbo",
34-
"litellm_params": {"model": "groq/llama3-8b-8192"},
34+
"litellm_params": {"model": "groq/llama-3.1-8b-instant"},
3535
"model_info": {"id": "groq-llama"},
3636
},
3737
]
@@ -182,7 +182,7 @@ async def test_get_available_endpoints_tpm_rpm_check_async(ans_rpm):
182182
},
183183
{
184184
"model_name": "gpt-3.5-turbo",
185-
"litellm_params": {"model": "groq/llama3-8b-8192"},
185+
"litellm_params": {"model": "groq/llama-3.1-8b-instant"},
186186
"model_info": {"id": "5678", "rpm": non_ans_rpm},
187187
},
188188
]

tests/local_testing/test_model_alias_map.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515

1616
litellm.set_verbose = True
1717

18-
model_alias_map = {"good-model": "groq/llama3-8b-8192"}
18+
model_alias_map = {"good-model": "groq/llama-3.1-8b-instant"}
1919

2020

2121
def test_model_alias_map(caplog):

tests/local_testing/test_router.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -119,7 +119,7 @@ async def test_router_provider_wildcard_routing():
119119
print("response 2 = ", response2)
120120

121121
response3 = await router.acompletion(
122-
model="groq/llama3-8b-8192",
122+
model="groq/llama-3.1-8b-instant",
123123
messages=[{"role": "user", "content": "hello"}],
124124
)
125125

tests/local_testing/test_router_batch_completion.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ async def test_batch_completion_multiple_models(mode):
4444
{
4545
"model_name": "groq-llama",
4646
"litellm_params": {
47-
"model": "groq/llama3-8b-8192",
47+
"model": "groq/llama-3.1-8b-instant",
4848
},
4949
},
5050
]
@@ -143,7 +143,7 @@ async def test_batch_completion_fastest_response_streaming():
143143
{
144144
"model_name": "groq-llama",
145145
"litellm_params": {
146-
"model": "groq/llama3-8b-8192",
146+
"model": "groq/llama-3.1-8b-instant",
147147
},
148148
},
149149
]
@@ -179,7 +179,7 @@ async def test_batch_completion_multiple_models_multiple_messages():
179179
{
180180
"model_name": "groq-llama",
181181
"litellm_params": {
182-
"model": "groq/llama3-8b-8192",
182+
"model": "groq/llama-3.1-8b-instant",
183183
},
184184
},
185185
]

tests/local_testing/test_spend_calculate_endpoint.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ async def test_spend_calc_model_on_router_messages():
4343
{
4444
"model_name": "special-llama-model",
4545
"litellm_params": {
46-
"model": "groq/llama3-8b-8192",
46+
"model": "groq/llama-3.1-8b-instant",
4747
},
4848
}
4949
]
@@ -86,7 +86,7 @@ async def test_spend_calc_using_response():
8686
}
8787
],
8888
"created": "1677652288",
89-
"model": "groq/llama3-8b-8192",
89+
"model": "groq/llama-3.1-8b-instant",
9090
"object": "chat.completion",
9191
"system_fingerprint": "fp_873a560973",
9292
"usage": {

tests/test_litellm/test_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1051,7 +1051,7 @@ def test_proxy_function_calling_support_consistency(
10511051
("litellm_proxy/claude-prod", "anthropic/claude-3-sonnet-20240229", False),
10521052
("litellm_proxy/claude-dev", "anthropic/claude-3-haiku-20240307", False),
10531053
# Groq with custom names (cannot be resolved)
1054-
("litellm_proxy/fast-llama", "groq/llama3-8b-8192", False),
1054+
("litellm_proxy/fast-llama", "groq/llama-3.1-8b-instant", False),
10551055
("litellm_proxy/groq-gemma", "groq/gemma-7b-it", False),
10561056
# Cohere with custom names (cannot be resolved)
10571057
("litellm_proxy/cohere-command", "cohere/command-r", False),

tests/test_openai_endpoints.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -550,7 +550,7 @@ async def test_proxy_all_models():
550550
async with aiohttp.ClientSession() as session:
551551
# call chat/completions with a model that the key was not created for + the model is not on the config.yaml
552552
await chat_completion(
553-
session=session, key=LITELLM_MASTER_KEY, model="groq/llama3-8b-8192"
553+
session=session, key=LITELLM_MASTER_KEY, model="groq/llama-3.1-8b-instant"
554554
)
555555

556556
await chat_completion(

0 commit comments

Comments
 (0)