Skip to content

Commit 36ea2b8

Browse files
authored
docs: Update image size
1 parent 041623b commit 36ea2b8

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

docs/source/get-started/adaptive-batching.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,14 +19,14 @@ Adaptive batching is implemented on the server side. This is advantageous as opp
1919
Specifically, there is a dispatcher within a BentoML Service that oversees collecting requests into a batch until the conditions of the batch window or batch size are met, at which point the batch is sent to the model for inference.
2020

2121
.. image:: ../../_static/img/get-started/adaptive-batching/single-service-batching.png
22-
:width: 75%
22+
:width: 55%
2323
:align: center
2424
:alt: Adaptive batching in a single BentoML Service
2525

2626
For multiple Services, the Service responsible for running model inference (``ServiceTwo`` in the diagram below) collects requests from the intermediary Service (``ServiceOne``) and forms batches based on optimal latency.
2727

2828
.. image:: ../../_static/img/get-started/adaptive-batching/multi-service-batching.png
29-
:width: 100%
29+
:width: 75%
3030
:align: center
3131
:alt: Adaptive batching in multiple BentoML Services
3232

0 commit comments

Comments
 (0)