Skip to content

Commit bdc5f06

Browse files
Fix broken modal.com URLs: blog typos and wrong doc paths (#1503)
Co-authored-by: Devin AI <158243242+devin-ai-integration[bot]@users.noreply.github.com>
1 parent 2146ea2 commit bdc5f06

File tree

5 files changed

+5
-5
lines changed

5 files changed

+5
-5
lines changed

06_gpu_and_ml/llm-serving/ministral3_inference.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@
7070

7171
# Native hardware support for FP8 formats in [Tensor Cores](https://modal.com/gpu-glossary/device-hardware/tensor-core)
7272
# is limited to the latest [Streaming Multiprocessor architectures](https://modal.com/gpu-glossary/device-hardware/streaming-multiprocessor-architecture),
73-
# like those of Modal's [Hopper H100/H200 and Blackwell B200 GPUs](https://modal.com/blog/announcing-h200-b200).
73+
# like those of Modal's [Hopper H100/H200 and Blackwell B200 GPUs](https://modal.com/blog/introducing-b200-h200).
7474

7575
# At 80 GB VRAM, a single H100 GPU has enough space to store the 8B FP8 model weights (~8 GB)
7676
# and a very large KV cache. A single H100 is also enough to serve the 14B model in full precision,

06_gpu_and_ml/llm-serving/very_large_models.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -277,7 +277,7 @@ def _start_server() -> subprocess.Popen:
277277
app = modal.App("example-serve-very-large-models", image=image)
278278

279279
# Most importantly, we need to decide what hardware to run on.
280-
# [H200 and B200 GPUs](https://modal.com/blog/introducting-b200-h200)
280+
# [H200 and B200 GPUs](https://modal.com/blog/introducing-b200-h200)
281281
# have over 100 GB of [GPU RAM](https://modal.com/gpu-glossary/device-hardware/gpu-ram) --
282282
# 141 GB and 180 GB, respectively.
283283
# The model's weights will be stored in this memory,

06_gpu_and_ml/llm-serving/vllm_inference.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@
5252
# We'll use an FP8 (eight-bit floating-point) post-training-quantized variant: `Qwen/Qwen3-4B-Thinking-2507-FP8`.
5353
# Native hardware support for FP8 formats in [Tensor Cores](https://modal.com/gpu-glossary/device-hardware/tensor-core)
5454
# is limited to the latest [Streaming Multiprocessor architectures](https://modal.com/gpu-glossary/device-hardware/streaming-multiprocessor-architecture),
55-
# like those of Modal's [Hopper H100/H200 and Blackwell B200 GPUs](https://modal.com/blog/announcing-h200-b200).
55+
# like those of Modal's [Hopper H100/H200 and Blackwell B200 GPUs](https://modal.com/blog/introducing-b200-h200).
5656

5757
# You can swap this model out for another by changing the strings below.
5858
# A single H100 GPU has enough VRAM to store a 4,000,000,000 parameter model,

06_gpu_and_ml/speech-to-text/sortformer2_1_speaker_diarization.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@
7474

7575
# ## Using WebSockets to stream audio and diarization results
7676

77-
# We use a Modal [ASGI](https://modal.com/docs/guide/asgi) app to serve the diarization results
77+
# We use a Modal [ASGI](https://modal.com/docs/guide/webhooks) app to serve the diarization results
7878
# over WebSockets. This allows us to stream the diarization results to the client in real-time.
7979

8080
# We use a simple queue-based architecture to handle the audio and diarization results.

13_sandboxes/sandbox_agent.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515

1616
app = modal.App.lookup("example-sandbox-agent", create_if_missing=True)
1717

18-
# First, we create a custom [Image](https://modal.com/docs/images) that has Claude Code
18+
# First, we create a custom [Image](https://modal.com/docs/guide/images) that has Claude Code
1919
# and git installed.
2020

2121
image = (

0 commit comments

Comments
 (0)