Skip to content
Discussion options

You must be logged in to vote

Thanks for the extra context. After some extensive testing, we were able to reproduce the same memory behaviour, but the potential causes for that do not seem to point to a memory leak. Let's move this to the discussion forum as the underlying issue is not a bug per-se.

Background

A bit of background on how the transformer pipeline works during inference: The user passes in strings or Doc instances to the model's pipe method. In the case of the former, the model initially runs the tokenizer on the strings and constructs Doc objects, since pipeline components only work with Doc inputs.

When a batch of documents are passed to the predict method of the Transformer pipe, it has to split the t…

Replies: 3 comments 3 replies

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
3 replies
@saketsharmabmb
Comment options

@saketsharmabmb
Comment options

@shadeMe
Comment options

Answer selected by adrianeboyd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
perf / memory Performance: memory use feat / transformer Feature: Transformer
2 participants
Converted from issue

This discussion was converted from issue #12037 on January 11, 2023 15:47.