Replies: 1 comment
-
KM supports "Request Context" parameters, which can be used to change configurations during a request, without affecting the deployment or other concurrent requests. To implement what you suggest, you could add a new Request Context param name, and support it in the Ollama generator. Here's some pointers:
To access params stored in the request context, you can inject |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello everyone,
I'm currently working on a project where I'd like to use multiple models in Ollama for Kernel Memory. Specifically, I want to be able to select which Ollama model is used when querying the Kernel Memory with ask. I added the kernel manager as a .net core web api into my dotnet aspire solution.
Is there a standard or recommended way to implement this type of functionality? Would it be necessary to write a custom MemoryService to enable model selection, or is there an existing feature or pattern that supports this?
Any guidance or examples from those who have faced a similar requirement would be greatly appreciated!
Thank you in advance!
Beta Was this translation helpful? Give feedback.
All reactions