Skip to content
Discussion options

You must be logged in to vote

Hey :)

so the RAM needed does not depend on the total length of the sequence as the sequence is broken into batches of size batch-size in the first place. This is done automatically and you don't need to split anything.
While a larger batch-size is faster, I don't think it has a significant effect in Inference mode, so I would not worry. If you use the GPU not only for DECODE but you have an actual screen attached, you might end up with varying numbers anyway. The only thing that can be tricky is if you really max out your GPU RAM utilisation with DECODE and then open a video game for example, cause that could then break the fitting.

batch_size='auto' was a bit fast, since you have seen t…

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@cleterrier
Comment options

@Haydnspass
Comment options

@Haydnspass
Comment options

Answer selected by cleterrier
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants