Skip to content

Commit fc0dba4

Browse files
committed
fix typo
Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>
1 parent 4aa5a6a commit fc0dba4

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/guides/sequential_onloading.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ Before a model can be sequentially onloaded, it must first be broken up into dis
2828
*This image depicts some of the operations performed when executing the Llama3.2-Vision model*
2929

3030
![sequential_onloading](../assets/sequential_decoder_layers.jpg)
31-
*This image depicts the sequential text decoder layers of the Llama3.2-Vision model. Each of the individual decoder layers*
31+
*This image depicts the sequential text decoder layers of the Llama3.2-Vision model. Each of the individual decoder layers is onloaded separately*
3232

3333
## Sequential Targets and Usage ##
3434
You can use sequential onloading by calling `oneshot` with the `pipeline="sequential"` argument. Note that this pipeline is the default for all oneshot calls which require calibration data. If the sequential pipeline proves to be problematic, you can specify `pipeline="basic"` to use a basic pipeline which does not require sequential onloading, but only works performantly when the model is small enough to fit into the available VRAM.

0 commit comments

Comments
 (0)