-
-
Notifications
You must be signed in to change notification settings - Fork 6
Open
Description
Use cases:
- You can fit the whole model into GPU ram
- You can fit part of the model into GPU ram
- You need keep all the model weights on disk
In all these cases, we should be able to detect how much GPU ram is available, and determine the max amount of model to store that way. More advanced use cases of sharing GPU with other applications may need manual control over the memory, but that can be done later.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels