Skip to content

Commit cfe4fe8

Browse files
committed
Merge branch 'mlp-tutorials-update-iii' of github.com:lukasgd/cscs-docs into mlp-tutorials-update-iii
2 parents c5383db + c1c72e7 commit cfe4fe8

File tree

4 files changed

+4
-4
lines changed

4 files changed

+4
-4
lines changed

docs/build-install/containers.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ Its command-line interface (CLI) closely mirrors Docker’s, providing a consist
77
[](){#ref-build-containers-configure-podman}
88
## Preliminary step: configuring Podman's storage
99

10-
The first step in order to use Podman on Alps is to create a valid Container Storage configuration file in your home according to the following minimal template:
10+
The first step in order to use Podman on Alps is to create a valid Container Storage configuration file in your home directory, according to the following minimal template:
1111

1212
```toml title="$HOME/.config/containers/storage.conf"
1313
[storage]

docs/software/ml/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
# Machine learning applications and frameworks
33

44
CSCS supports a wide range of machine learning (ML) applications and frameworks on its systems.
5-
Most ML workloads are containerized to ensure portability, reproducibility, and ease of use across machines.
5+
Most ML workloads are containerized to ensure portability, reproducibility, and ease of use across systems.
66

77
Users can choose between running containers, using provided uenv software stacks, or building custom Python environments tailored to their needs.
88

docs/software/ml/pytorch.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ user@nidYYYYYY$ source venv-ngc-pt-25.06/bin/activate # (3)!
126126
The changes made to the virtual environment will outlive the container as they are persisted on the distributed filesystem.
127127

128128
!!! note
129-
Keep in mind that
129+
Keep in mind that:
130130

131131
* this virtual environment is _specific_ to this particular container and won't actually work unless you are using it from inside this container - it relies on the resources packaged inside the container.
132132
* every Slurm job making use of this virtual environment will need to activate it first (_inside_ the `srun` command).

docs/tutorials/ml/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
The LLM tutorials gradually introduce key concepts of the Machine Learning Platform in a series of hands-on examples. A particular focus is on the [Container Engine][ref-container-engine] for managing the runtime environment.
55

6-
In a [first tutorial][software-ml-llm-inference-tutorial], you will learn how to run inference with a LLM on a single node using a container from the NVIDIA GPU Cloud (NGC). Concepts such as container environment description, layering a thin virtual environment on top of the container image, and job launching and monitoring will be introduced.
6+
In the [first tutorial][software-ml-llm-inference-tutorial], you will learn how to run inference with a LLM on a single node using a container from the NVIDIA GPU Cloud (NGC). Concepts such as container environment description, layering a thin virtual environment on top of the container image, and job launching/monitoring will be introduced.
77

88
Building on the first tutorial, in the [second tutorial][software-ml-llm-fine-tuning-tutorial] you will learn how to train (fine-tune) a LLM on multiple GPUs on a single node. For this purpose, you will use HuggingFace's `accelerate` and see best practices for dataset management.
99

0 commit comments

Comments
 (0)