Conversation memory summary after returning LLM response to user #11345
Replies: 1 comment
-
🤖 Hi there, Thank you for your insightful question. You're correct in your understanding of the current flow of the LangChain framework. The To answer your questions:
That being said, your suggestion to improve user experience by reducing the wait time for the response is valid and worth considering. One possible solution could be to make the memory update operation asynchronous, so that it can be performed in the background after the response is returned to the user. However, this would require careful design to avoid the potential issues I mentioned above. I hope this answers your questions. If you have any further questions or if anything is unclear, please let me know! Best, SourcesThis response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research. Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant. If you want to continue the conversation, start your reply with @dosu-bot. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all, I have an inquiry regarding the conversation history with langchain.
Currently from what I saw, the LLM chain tries to store the conversation memory before returning the LLM output, and in case of any conversation summary memory was taken into place, this means that before we return the response to the user (let's say in a chat), there is an additional call to the model to construct the summary of the conversation so far.
This means that we are making the user pend for no reason as the flow can simply be:
Which will result in a better user experience, in my opinion.
So my questions here are:
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions