Skip to content

FP16 does not decrease much GPU memory #13252

Discussion options

You must be logged in to vote

Hi @icoz69 AFAIK, the memory usage depends on your model architecture, specifically, the ratio of the model size to the size of activations. This is because, with amp, your model always stays in fp32 while some operations in your model are done in fp16.

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@m13uz
Comment options

@akihironitta
Comment options

Answer selected by akihironitta
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment