Skip to content

Commit 63e38a1

Browse files
authored
slurm -> Slurm (#175)
1 parent 87bd8ff commit 63e38a1

File tree

6 files changed

+12
-12
lines changed

6 files changed

+12
-12
lines changed

docs/clusters/bristen.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Bristen consists of 32 A100 nodes [NVIDIA A100 nodes][ref-alps-a100-node]. The n
1212
|-----------|--------| ----------------- | ---------- |
1313
| [a100][ref-alps-a100-node] | 32 | 32 | 128 |
1414

15-
Nodes are in the [`normal` slurm partition][ref-slurm-partition-normal].
15+
Nodes are in the [`normal` Slurm partition][ref-slurm-partition-normal].
1616

1717
### Storage and file systems
1818

@@ -48,7 +48,7 @@ Users are encouraged to use containers on Bristen.
4848

4949
Bristen uses [Slurm][ref-slurm] as the workload manager, which is used to launch and monitor distributed workloads, such as training runs.
5050

51-
There is currently a single slurm partition on the system:
51+
There is currently a single Slurm partition on the system:
5252

5353
* the `normal` partition is for all production workloads.
5454
+ nodes in this partition are not shared.

docs/clusters/clariden.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ The number of nodes can change when nodes are added or removed from other cluste
1414
|-----------|--------| ----------------- | ---------- |
1515
| [gh200][ref-alps-gh200-node] | 1,200 | 4,800 | 4,800 |
1616

17-
Most nodes are in the [`normal` slurm partition][ref-slurm-partition-normal], while a few nodes are in the [`debug` partition][ref-slurm-partition-debug].
17+
Most nodes are in the [`normal` Slurm partition][ref-slurm-partition-normal], while a few nodes are in the [`debug` partition][ref-slurm-partition-debug].
1818

1919
### Storage and file systems
2020

@@ -71,7 +71,7 @@ Alternatively, [uenv][ref-uenv] are also available on Clariden. Currently deploy
7171

7272
Clariden uses [Slurm][ref-slurm] as the workload manager, which is used to launch and monitor distributed workloads, such as training runs.
7373

74-
There are two slurm partitions on the system:
74+
There are two Slurm partitions on the system:
7575

7676
* the `normal` partition is for all production workloads.
7777
* the `debug` partition can be used to access a small allocation for up to 30 minutes for debugging and testing purposes.

docs/running/jobreport.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ The report is divided into two parts: a general summary and GPU specific values.
5656
| Field | Description |
5757
| ----- | ----------- |
5858
| Job Id | The Slurm job id |
59-
| Step Id | The slurm step id. A job step in Slurm is a subdivision of a job started with srun |
59+
| Step Id | The Slurm step id. A job step in Slurm is a subdivision of a job started with srun |
6060
| User | The user account that submitted the job |
6161
| Slurm Account | The project account that will be billed |
6262
| Start Time, End Time, Elapsed Time | The time the job started and ended, and how long it ran |
@@ -77,7 +77,7 @@ The report is divided into two parts: a general summary and GPU specific values.
7777
| SM Utilization % | The percentage of the process's lifetime during which Streaming Multiprocessors (SM) were executing a kernel |
7878
| Memory Utilization % | The percentage of process's lifetime during which global (device) memory was being read or written |
7979

80-
## Example with slurm: srun
80+
## Example with Slurm: srun
8181

8282
The simplest example to test `jobreport` is to run it with the sleep command.
8383
It is important to separate `jobreport` (and its options) and your command with `--`.
@@ -155,7 +155,7 @@ GPU Specific Values
155155
4. Uncheck "Set locale environment variables on startup"
156156
5. Quit and reopen the terminal and try again. This should fix the issue.
157157

158-
## Example with slurm: batch script
158+
## Example with Slurm: batch script
159159

160160
The `jobreport` command can be used in a batch script
161161
The report printing, too, can be included in the script and does not need the `srun` command.

docs/software/communication/openmpi.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ OpenMPI is provided through a [uenv][ref-uenv] similar to [`prgenv-gnu`][ref-uen
2222
Once the uenv is loaded, compiling and linking with OpenMPI and libfabric is transparent.
2323
At runtime, some additional options must be set to correctly use the Slingshot network.
2424

25-
First, when launching applications through slurm, [PMIx](https://pmix.github.com) must be used for application launching.
25+
First, when launching applications through Slurm, [PMIx](https://pmix.github.com) must be used for application launching.
2626
This is done with the `--mpi` flag of `srun`:
2727
```bash
2828
srun --mpi=pmix ...

docs/software/sciapps/cp2k.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ On our systems, CP2K is built with the following dependencies:
6565

6666
### Running on the HPC platform
6767

68-
To start a job, two bash scripts are potentially required: a [slurm] submission script, and a wrapper to start the [CUDA
68+
To start a job, two bash scripts are potentially required: a [Slurm] submission script, and a wrapper to start the [CUDA
6969
MPS] daemon so that multiple MPI ranks can use the same GPU.
7070

7171
```bash title="run_cp2k.sh"
@@ -138,7 +138,7 @@ sbatch run_cp2k.sh
138138

139139
Each GH200 node has 4 modules, each of them composed of a ARM Grace CPU with 72 cores and a H200 GPU directly
140140
attached to it. Please see [Alps hardware][ref-alps-hardware] for more information.
141-
It is important that the number of MPI ranks passed to [slurm] with `--ntasks-per-node` is a multiple of 4.
141+
It is important that the number of MPI ranks passed to [Slurm] with `--ntasks-per-node` is a multiple of 4.
142142

143143
??? note
144144

@@ -524,5 +524,5 @@ As a workaround, you can disable CUDA acceleration for the grid backend:
524524
[OpenBLAS]: http://www.openmathlib.org/OpenBLAS/
525525
[Intel MKL]: https://www.intel.com/content/www/us/en/developer/tools/oneapi/onemkl.html
526526
[Cray MPICH]: https://docs.nersc.gov/development/programming-models/mpi/cray-mpich/
527-
[slurm]: https://slurm.schedmd.com/
527+
[Slurm]: https://slurm.schedmd.com/
528528
[CUDA MPS]: https://docs.nvidia.com/deploy/mps/index.html

docs/software/uenv/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -213,7 +213,7 @@ This is very useful for interactive sessions, for example if you want to work in
213213
$ make -j
214214

215215
# run the affinity executable on two nodes - note how the uenv is
216-
# automatically loaded by slurm on the compute nodes, because CUDA and MPI from
216+
# automatically loaded by Slurm on the compute nodes, because CUDA and MPI from
217217
# the uenv are required to run.
218218
$ srun -n2 -N2 ./affinity.cuda
219219
GPU affinity test for 2 MPI ranks

0 commit comments

Comments
 (0)