Dynamic Top-k Retrieval Chunks #5094
Unanswered
Malak-Alaabiad
asked this question in
Q&A
Replies: 1 comment
-
this sounds like ProblemMap No.10 (paired mapping / format drift). |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Suggest me a specific node or flow to reduce the number of tokens going into the LLM model, considering that my data is stored in a Qdrant collection, and I'm using a custom retriever node to pull only the necessary metadata. This custom retriever node is connected to the Conversational Retriever QA Chain, which then passes the data directly to the LLM.
Now, I want to implement a Dynamic Top-k Retrieval Chunks or a similar flow to achieve the same goal—reducing the tokens sent to the model, which would help minimize the associated costs.
Beta Was this translation helpful? Give feedback.
All reactions