-
There's any way to keep a model loaded when using the api? Example, on a first request the model "sdxl.safetensors" is loaded, but on the second request it got loaded again, slowing the api |
Beta Was this translation helpful? Give feedback.
Answered by
asagi4
Mar 24, 2024
Replies: 1 comment 6 replies
-
What kind of workflows are you sending to the API? In my experience the model stays loaded as long as the prompt you send in doesn't change in ways that require reloading. Are you using the same node ID for the checkpoint loader between prompts? |
Beta Was this translation helpful? Give feedback.
6 replies
Answer selected by
nephi-dev
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
What kind of workflows are you sending to the API? In my experience the model stays loaded as long as the prompt you send in doesn't change in ways that require reloading.
Are you using the same node ID for the checkpoint loader between prompts?