-
Notifications
You must be signed in to change notification settings - Fork 7
Open
Labels
bugSomething isn't workingSomething isn't working
Description
Describe the bug
The batch size used for each consumer is determined by the batch size of the model run by the consumer. The GrpcModelWrapper.get_batch_size in grpc_clients.py to determine this. The maximum batch size for a consumer is determined with the environment variable TF_MAX_BATCH_SIZE in the helmfile for each consumer. If this value is larger than the maximum batch size set by the environment variable MAX_BATCH_SIZE in the tf-serving helmfile, the prediction will fail.
To Reproduce
Steps to reproduce the behavior:
- Set
TF_MAX_BATCH_SIZEin the helmfile for the consumer to a number larger thanMAX_BATCH_SIZEin the tf-serving helmfile. For example, for the segmentation consumer, setTF_MAX_BATCH_SIZEin the segmentation consumer helmfile to 64 and ensure thatMAX_BATCH_SIZEin the tf-serving helmfile is 32. - Upload an image to the kiosk for prediction.
- See error: "Task size 64 is larger than maximum input batch size 32"
Expected behavior
Enforce that the maximum batch size of the consumer is smaller than the maximum batch size of tf-serving before prediction.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working