Skip to content

Commit 5663536

Browse files
Update README.md
1 parent 93732ef commit 5663536

File tree

1 file changed

+2
-2
lines changed
  • cloud-infrastructure/ai-infra-gpu/ai-infrastructure/dstack

1 file changed

+2
-2
lines changed

cloud-infrastructure/ai-infra-gpu/ai-infrastructure/dstack/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -121,7 +121,7 @@ fsdp_config:
121121
* `distributed_type`: `FSDP` indicates the use of Fully Sharded Data Parallel (FSDP), a technique that enables training large models that would otherwise not fit on a single GPU.
122122
* `fsdp_config`: These set up how FSDP operates, such as how the model is sharded (`fsdp_sharding_strategy`) and whether parameters are offloaded to CPU (`fsdp_offload_params`).
123123

124-
![Hybrid shards](cloud-infrastructure/ai-infra-gpu/ai-infrastructure/dstack/assets/images/image2.png "Hybrid shards")
124+
![Hybrid shards](assets/images/image2.png "Hybrid shards")
125125

126126
With the `FSDP` of `distributed_type` and `FULL_SHARD` of `fsdp_config`’s `fsdp_sharding_strategy`, a model will be sharded across multiple GPUs in a single machine. When dealing with multiple compute nodes, each node will host an identical copy of the model, which is itself split across multiple GPUs within that node. This means each partitioned model instance on each node processes different sections or batches of your dataset. To distribute a single model across multiple GPUs spanning across multiple nodes, configure the parameter `fsdp_sharding_strategy` as `HYBRID_SHARD`.
127127

@@ -208,7 +208,7 @@ curl -X POST https://black-octopus-1.mycustomdomain.com/generate \
208208

209209
Additionally, for a deployed model, dstack automatically provides a user interface to directly interact with the model:
210210

211-
![User interface](cloud-infrastructure/ai-infra-gpu/ai-infrastructure/dstack/assets/images/image1.png "User interface")
211+
![User interface](assets/images/image1.png "User interface")
212212

213213
## Conclusion
214214

0 commit comments

Comments
 (0)