[Feature] Machine Learning should unload model in case of GPU memory issues #18131
schuettecarsten
started this conversation in
Feature Request
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I have searched the existing feature requests, both open and closed, to make sure this is not a duplicate request.
The feature
My GPU has 4 GB, which is not enough for all models. Immich ML unloads models after some time, but if new images are uploaded, Smart Search and Face Recognition run at the same time, loading their models. That sometimes leads to OOM errors. It is somewhat unclear when Immich retries to process the images or if they are simply ignored in Smart Search and Face Recognition if the ML job fails. Also, it would be great if Immich ML could prefer pending jobs for already unloaded models and - if there is not enough memory to load a model, first try to unload other models and try again.
The exceptions are:
They are expected and fine, so it is not really a bug. But the way how ML module behaves in such cases can be optimized.
Platform
Beta Was this translation helpful? Give feedback.
All reactions