modelsio
providers for models
resource
#1528
Replies: 2 comments 2 replies
-
Can /files api https://github.com/meta-llama/llama-stack/blob/main/llama_stack/apis/files/files.py fulfill the modelsio job here? |
Beta Was this translation helpful? Give feedback.
-
one part of this that I think will be tricky is:
which in this proposal seems to be a model retriever. But won't we also then want models.register to modify, say, an inference provider to use this new model? If we go down the route of "registration" being a downloader in a way, we might not want to overload it and also make a new API for plopping a new model at an inference endpoint. I could also be missing something here |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Llama Stack has prioritized Llama models for good reasons, but recent contributions have seen the project go the direction of more model neutrality.
Users may want to bring model parameters from cloud object storage:
and use those modules uniformly across LLS.
modelsio
providers would expose a few endpoints so they can bridge to remote providers as well:The abstraction offered by inline vs. remote providers breaks down a bit when plumbing large artifacts around, but this way, we can implement the following user behavior:
curl -X 'POST' -F "./my-hf-model.tar" -d '{"unique_model_uri": "localfs-huggingface-myusername-my-hf-model"}' https://myllamastack/api/modelsio/localfs-huggingface/tar
curl -X 'POST' -d '{"provider_id": "localfs-huggingface", "model_id": "localfs-huggingface-myusername-my-hf-model"} https://myllamastack/api/models/'
curl -X 'POST' -d '{...}' https://myllamastack/api/evaluation/<evaluation_type>
The example workflow is pretty far from the existing APIs but explains the flow.
Gaps
Beta Was this translation helpful? Give feedback.
All reactions