Llama-3.1 405b with litellm proxy has issues #3880
Replies: 3 comments
-
Check your terminal logs and/or the logs directory at project root. |
Beta Was this translation helpful? Give feedback.
-
Hi, Running litellm locally with detailed log enabled. This is what we get on Librechat logs: 2024-09-03 22:11:53 error: [handleAbortError] AI response error; aborting request: litellm.APIConnectionError: Error parsing chunk: Expecting property name enclosed in double quotes: line 1 column 3 (char 2), During handling of the above exception, another exception occurred: Traceback (most recent call last): And this is what the logs at litellm side gives: 22:11:42 - LiteLLM Proxy:DEBUG: parallel_request_limiter.py:28 - Inside Max Parallel Request Pre-Call Hook 22:11:42 - LiteLLM:DEBUG: utils.py:247 - Request to litellm: 22:11:42 - LiteLLM:DEBUG: utils.py:247 - ASYNC kwargs[caching]: False; litellm.cache: None; kwargs.get('cache'): None POST Request Sent from LiteLLM: 22:11:50 - LiteLLM:DEBUG: main.py:5133 - makes async anthropic streaming POST request 22:11:52 - LiteLLM:DEBUG: utils.py:247 - Logging Details LiteLLM-Success Call: Cache_hit=False 22:11:53 - LiteLLM:DEBUG: utils.py:247 - Token Counter - using generic token counter, for model=meta/llama3-405b-instruct-maas 22:11:53 - LiteLLM:DEBUG: caching.py:33 - InMemoryCache: set_cache. current size= 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): During handling of the above exception, another exception occurred: Traceback (most recent call last): During handling of the above exception, another exception occurred: Traceback (most recent call last): During handling of the above exception, another exception occurred: Traceback (most recent call last): 22:11:53 - LiteLLM Proxy:DEBUG: proxy_server.py:2596 - An error occurred: litellm.APIConnectionError: Error parsing chunk: Expecting property name enclosed in double quotes: line 1 column 3 (char 2), During handling of the above exception, another exception occurred: Traceback (most recent call last): Debug this by setting |
Beta Was this translation helpful? Give feedback.
-
Hi @ss-gonda , thanks for using LiteLLM. Any chance we can hop on a call to learn how we can improve LiteLLM Proxy for you ? We’re planning roadmap and I’d love to get your feedback my cal for your convenience: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
What happened?
Steps to Reproduce
What browsers are you seeing the problem on?
No response
Relevant log output
Based on log analysis on litellm and librechat. I can say that litellm latest complains about unsuppored params: presence_penalty frequency_penalty I was able to fix by some code changes and removing these params (not well structured). After that it the chat works fine but does not gives chattitles again because of above parameters. Also, after all the fixes it keeps giving "assistant" randomly at start of response but it looks like an issue with Llama deployed model.
Screenshots
No response
Code of Conduct
Beta Was this translation helpful? Give feedback.
All reactions