You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/navigating_the_cluster/hpc_foundations.md
+5-6Lines changed: 5 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ sidebar_position: 2
7
7
8
8
The goal of this exercise is to help you understand the fundamentals **_A to Z_** on effecively navigating the cluster for your research or academic projects.
9
9
10
-
Before we begin this exercise please make sure you have access to the NYU HPC cluster, if not please review the \[Accessing HPC page].
10
+
Before we begin this exercise please make sure you have access to the NYU HPC cluster, if not please review the [Accessing HPC page](../accessing_hpc/accessing_hpc.md).
11
11
12
12
Login to the Greene cluster with ssh at :
13
13
> Accessible under NYU Net ( either via VPN or within campus network )
@@ -120,7 +120,7 @@ Similar to `/home`, users have access to multiple filesystems that are :
120
120
| /scratch | /scratch/**Net_ID**/ | General Storage | $SCRATCH
You will find more details about these filesystems at \[Greene Storage Types page].
123
+
You will find more details about these filesystems at [Greene Storage Types page](../storage_specs.mdx).
124
124
125
125
You can jump to your `/scratch` directory at `/scratch/Net_ID/` with the `cd` command as `cd /scratch/Net_ID`, Or you could simple use the `$SCRATCH` environment variable as:
126
126
@@ -204,6 +204,7 @@ print("hello, world")
204
204
[pp2959@log-3 ~]$
205
205
```
206
206
207
+
207
208
> Here `os.execute()` executes a shell command, in this example the command `hostname` to print the name of the host on which the script is being executed. Followed by printing the message `hello, world`
208
209
209
210
Now if you try to run the script as `lua hello.lua`, you may get an error like:
@@ -423,8 +424,6 @@ hello, world
423
424
>
424
425
> 4. Based on your output, you may notice the name of the compute node that this program runs on, the node `cm001.hpc.nyu.ed` in this example is a CPU only node, you may notice a different node. You can find more details about the \[specific nodes here].
425
426
426
-
<br/>
427
-
428
427
**_Now how do we determine Or specify the amount of resources needed to run our `hello.lua` script ?_**
429
428
430
429
By defualt slurm schedules just **_1 CPU_** and **_1 GB_** memory to run your programs.
@@ -759,6 +758,7 @@ srun: job 56142474 has been allocated resources
759
758
>
760
759
> - `--label` labels standard output of tasks based on task ID from 0 to N.
761
760
761
+
762
762
So far we understood that slurm chooses '_on which nodes our programs should run on_' based on it's scheduling decisions, however it also provides more control like specifying explicitly on which `partition` we can run our programs on.
763
763
764
764
Here partitions are similar nodes grouped together as a list. For example H100 nodes are grouped together as a partition named `H100_Partition`. Whenever we submit a job request for H100s then nodes sequentially along this partition are reserved and our job is scheduled on them.
@@ -775,8 +775,6 @@ To specify a particular partition, you can use the `--partition` option as shown
0 commit comments