Skip to content

Commit 1a063bf

Browse files
authored
Fixed the Acrolinx score
1 parent e1d6d33 commit 1a063bf

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

articles/cyclecloud/slurm.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ Slurm is a highly configurable open source workload manager. For more informatio
1616
> Starting with CycleCloud 8.4.0, the Slurm integration was rewritten to support new features and functionality. For more information, see [Slurm 3.0](slurm-3.md) documentation.
1717
1818
::: moniker range="=cyclecloud-7"
19-
Slurm can easily be enabled on a CycleCloud cluster by modifying the "run_list", in the configuration section of your cluster definition. A Slurm cluster has two main parts: the master (or scheduler) node, which runs the Slurm software on a shared file system, and the execute nodes, which mount that file system and run the submitted jobs. For example, a simple cluster template snippet may look like:
19+
To enable Slurm on a CycleCloud cluster, modify the "run_list" in the definiton of cluster's configuration section. A Slurm cluster has two main parts: the master (or scheduler) node, which runs the Slurm software on a shared file system, and the execute nodes, which mount that file system and run the submitted jobs. For example, a simple cluster template snippet may look like:
2020

2121
``` ini
2222
[cluster custom-slurm]
@@ -78,7 +78,7 @@ Slurm can easily be enabled on a CycleCloud cluster by modifying the "run_list"
7878
::: moniker-end
7979
## Editing Existing Slurm Clusters
8080

81-
Slurm clusters running in CycleCloud versions 7.8 and later implement an updated version of the autoscaling APIs that allows the clusters to utilize multiple nodearrays and partitions. To facilitate this functionality in Slurm, CycleCloud prepopulates the execute nodes in the cluster. Because of the prepopulation, you need to run a command on the Slurm scheduler node after making any changes to the cluster, such as autoscale limits or VM types.
81+
Slurm clusters running in CycleCloud versions 7.8 and later implement an updated version of the autoscaling APIs that allows the clusters to utilize multiple nodearrays and partitions. To facilitate this functionality in Slurm, CycleCloud prepopulates the executed nodes in the cluster. Because of the prepopulation, you need to run a command on the Slurm scheduler node after making any changes to the cluster, such as autoscale limits or VM types.
8282

8383
### Making Cluster Changes
8484

@@ -195,7 +195,7 @@ To override the UID and GID, click the edit button for both the `scheduler` node
195195
And the `execute` nodearray:
196196
![Edit Nodearray](~/articles/cyclecloud/images/slurmnodearraytab.png "Edit nodearray")
197197

198-
and add the next attributes to the `Configuration` section:
198+
Add the next attributes to the `Configuration` section:
199199
![Edit Configuration](~/articles/cyclecloud/images/slurmnodearrayedit.png "Edit configuration")
200200

201201
``` ini
@@ -215,22 +215,22 @@ CycleCloud uses Slurm's [Elastic Computing](https://slurm.schedmd.com/elastic_co
215215
[2019-12-09T21:19:03.400] power_save: pid 8629 waking nodes htc-1
216216
```
217217

218-
The other log to check is `/var/log/slurmctld/resume.log`. If the resume step is failing, there is `/var/log/slurmctld/resume_fail.log`. If there're messages about unknown or invalid node names, make sure you haven't added nodes to the cluster without next the steps in the "Making Cluster Changes" section above.
218+
The other log to check is `/var/log/slurmctld/resume.log`. If the resume step is failing, there's `/var/log/slurmctld/resume_fail.log`. If there're messages about unknown or invalid node names, make sure you haven't added nodes to the cluster without next the steps in the "Making Cluster Changes" section above.
219219

220220
## Slurm Configuration Reference
221221

222222
The next are the Slurm specific configuration options you can toggle to customize functionality:
223223

224224
| Slurm Specific Configuration Options | Description |
225225
| ------------------------------------ | ----------- |
226-
| slurm.version | Default: '18.08.7-1'. The Slurm version to install and run. This is currently the default and *only* option. In the future more versions of the Slurm software may be supported. |
226+
| slurm.version | Default: '18.08.7-1'. This sets the Slurm version to install and run. Right now, it’s the default and *only* option. More versions may be supported in the future. |
227227
| slurm.autoscale | Default: 'false'. A per-nodearray setting that controls whether Slurm should automatically stop and start nodes in this nodearray. |
228-
| slurm.hpc | Default: 'true'.A per-nodearray setting that controls whether nodes in the nodearray will be placed in the same placement group. Primarily used for nodearrays using VM families with InfiniBand. It only applies when slurm.autoscale is set to 'true'. |
228+
| slurm.hpc | Default: 'true'.A per-nodearray setting that controls whether nodes in the nodearray are placed in the same placement group. Primarily used for nodearrays using VM families with InfiniBand. It only applies when slurm.autoscale is set to 'true'. |
229229
| slurm.default_partition | Default: 'false'. A per-nodearray setting that controls whether the nodearray should be the default partition for jobs that don't request a partition explicitly. |
230230
| slurm.dampen_memory | Default: '5'. The percentage of memory to hold back for OS/VM overhead. |
231231
| slurm.suspend_timeout | Default: '600'. The amount of time (in seconds) between a suspend call and when that node can be used again. |
232232
| slurm.resume_timeout | Default: '1800'. The amount of time (in seconds) to wait for a node to successfully boot. |
233-
| slurm.install | Default: 'true'. Determines if the Slurm is installed at node boot ('true'). If Slurm is installed in a custom image this should be set to 'false' (proj version 2.5.0+). |
233+
| slurm.install | Default: 'true'. Determines if the Slurm is installed at node boot ('true'). If Slurm is installed in a custom image, this configuration option should be set to 'false' (proj version 2.5.0+). |
234234
| slurm.use_pcpu | Default: 'true'. A per-nodearray setting to control scheduling with hyperthreaded vcpus. Set to 'false' to set CPUs=vcpus in cyclecloud.conf. |
235235
| slurm.user.name | Default: 'slurm'. The username for the Slurm service to use. |
236236
| slurm.user.uid | Default: '11100'. The User ID to use for the Slurm user. |

0 commit comments

Comments
 (0)