Checking cache on model group fallbacks #4212
Manouchehri
started this conversation in
General
Replies: 1 comment
-
it seems reasonable to check cache when the cache key changes we have optimizations around reducing the number of cache get requests - e.g. batching gets for a key - https://docs.litellm.ai/docs/proxy/caching#turn-on-batch_redis_requests -- i don't view this as a bug, but i'm happy to convert to discussion and see what works best! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
What happened?
When a call in an LLM group fails, it seems like the cache is checked again. This is pretty wasteful since it's highly unlikely that this will ever result in a cache hit.
I would assume this is a bug across all caching providers, not just S3.
Relevant log output
No response
Twitter / LinkedIn details
https://www.linkedin.com/in/davidmanouchehri/
Beta Was this translation helpful? Give feedback.
All reactions