Skip to content

Commit 43c569d

Browse files
authored
Merge pull request #764 from wdpypere/infra_page
port ugent infra page to user docs
2 parents 76a8634 + f8e2f26 commit 43c569d

File tree

2 files changed

+97
-1
lines changed

2 files changed

+97
-1
lines changed

mkdocs/docs/HPC/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ Use the OS dropdown in the top bar to switch to a different operating system.
2020
{%- if site == 'Gent' %}
2121
- [Recording of HPC-UGent intro](https://www.ugent.be/hpc/en/training/introhpcugent-recording)
2222
- [Linux Tutorial](linux-tutorial/index.md)
23-
- [Hardware overview](https://www.ugent.be/hpc/en/infrastructure)
23+
- [Hardware overview](infrastructure.md)
2424
- [Available software](./only/gent/available_software/index.md)
2525
- [Migration of cluster and login nodes to RHEL9 (starting Sept'24)](rhel9.md)
2626
{%- endif %}

mkdocs/docs/HPC/infrastructure.md

Lines changed: 96 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,96 @@
1+
# Infrastructure
2+
3+
## Tier2 clusters of Ghent University
4+
5+
The Stevin computing infrastructure consists of several Tier2 clusters
6+
which are hosted in the S10 datacenter of Ghent University.
7+
8+
This infrastructure is co-financed by FWO and Department of Economy,
9+
Science and Innovation (EWI).
10+
11+
## Tier-2 login nodes
12+
13+
Log in to the HPC-UGent Tier-2 infrastructure via [https://login.hpc.ugent.be](https://login.hpc.ugent.be)
14+
or using SSH via `login.hpc.ugent.be`.
15+
16+
more info on using the web portal you can find [here](web_portal),
17+
and about connection with SSH [here](connecting).
18+
19+
## Tier-2 compute clusters
20+
21+
### CPU clusters
22+
23+
The HPC-UGent Tier-2 infrastructure currently included several standard
24+
CPU-only clusters, of different generations (listed from old to new).
25+
26+
For basic information on using these clusters, see our
27+
[documentation](running_batch_jobs.md).
28+
29+
| ***cluster name*** | ***# nodes*** | ***Processor architecture*** | ***Usable memory/node*** | ***Local diskspace/node*** | ***Interconnect*** | ***Operating system*** |
30+
| --- | --- | --- | --- | --- | --- | --- |
31+
| skitty | 72 | 2 x 18-core Intel Xeon Gold 6140 (Skylake @ 2.3 GHz) | 177 GiB | 1 TB & 240 GB SSD | EDR InfiniBand | RHEL 9 |
32+
| doduo (default cluster) | 128 | 2x 48-core AMD EPYC 7552 (Rome @ 2.2 GHz) | 250 GiB | 180GB SSD | HDR-100 InfiniBand | RHEL 8 |
33+
| gallade (*) | 16 | 2x 64-core AMD EPYC 7773X (Milan-X @ 2.2 GHz) | 940 GiB | 1.5 TB NVME | HDR-100 InfiniBand | RHEL 9 |
34+
| shinx | 48 | 2x 96-core AMD EPYC 9654 (Genoa @ 2.4 GHz) | 370 GiB | 500GB NVME | NDR-200 InfiniBand | RHEL 9 |
35+
36+
(*) also see this [extra information](./only/gent/2023/donphan-gallade#gallade-large-memory-cluster)
37+
38+
### Interactive debug cluster
39+
40+
41+
A special-purpose interactive debug cluster is available,
42+
where you should always be able to get a job running quickly,
43+
**without waiting in the queue**.
44+
45+
Intended usage is mainly for interactive work,
46+
either via an interactive job or using the [HPC-UGent web portal](web_portal).
47+
48+
This cluster is heavily over-provisioned, so jobs may
49+
run slower if the cluster is used more heavily.
50+
51+
Strict limits are in place per user:
52+
* max. 5 jobs in queue
53+
* max. 3 jobs running
54+
* max. of 8 cores and 27GB of memory in total for running jobs
55+
56+
For more information, see our [documentation](interactive_gent).
57+
58+
| ***cluster name*** | ***# nodes*** | ***Processor architecture*** | ***Usable memory/node*** | ***Local diskspace/node*** | ***Interconnect*** | ***Operating system*** |
59+
| --- | --- | --- | --- | --- | --- | --- |
60+
| donphan (*) | 16 | 2 x 18-core Intel Xeon Gold 6240 (Cascade Lake @ 2.6 GHz) + 1x shared NVIDIA Ampere A2 GPU (16GB GPU memory) | 738 GiB | 1.6 TB NVME | HDR-100 Infiniband | RHEL 8 |
61+
62+
(*) also see this [extra information](./only/gent/2023/donphan-gallade#donphan-debuginteractive-cluster)
63+
64+
### GPU clusters
65+
66+
GPU clusters are available in the HPC-UGent Tier-2 infrastructure,
67+
with different generations of NVIDIA GPUs.
68+
69+
These are well suited for specific workloads, with software that
70+
can leverage the GPU resources (like TensorFlow, PyTorch, GROMACS, AlphaFold, etc.).
71+
72+
For more information on using these clusters, see our documentation.
73+
74+
| ***cluster name*** | ***# nodes*** | ***Processor architecture & GPUs*** | ***Usable memory/node*** | ***Local diskspace/node*** | ***Interconnect*** | ***Operating system*** |
75+
| --- | --- | --- | --- | --- | --- | --- |
76+
| joltik | 10 | 2x 16-core Intel Xeon Gold 6242 (Cascade Lake @ 2.8 GHz) + 4x NVIDIA Volta V100 GPUs (32GB GPU memory) | 256 GiB | 800GB SSD | double EDR Infiniband | RHEL 9 |
77+
| accelgor | 9 | 2x 24-core AMD EPYC 7413 (Milan @ 2.2 GHz) + 4x NVIDIA Ampere A100 GPUs (80GB GPU memory) | 500 GiB | 180GB SSD | HDR InfiniBand | RHEL 8 |
78+
79+
80+
## Tier-2 shared storage
81+
82+
| ***Filesystem name*** | ***Intended usage*** | ***Total storage space*** | ***Personal storage space*** | ***VO storage space (^)*** |
83+
| ---| --- |---| --- | --- |
84+
| $VSC_HOME | Home directory, entry point to the system | 90 TB | 3GB (fixed) | (none) |
85+
| $VSC_DATA | Long-term storage of large data files | 1.9 PB | 25GB (fixed) | 250GB |
86+
| $VSC_SCRATCH | Temporary fast storage of 'live' data for calculations | 1.7 PB | 25GB (fixed) | 250GB |
87+
| $VSC_SCRATCH_ARCANINE | Temporary very fast storage of 'live' data for calculations (recommended for very I/O-intensive jobs) | 70 TB NVME | (none) | upon request |
88+
89+
90+
^ Storage space for a group of users (Virtual Organisation or VO for short) can be
91+
increased significantly on request. For more information, see our
92+
[documentation](running_jobs_with_input_output_data#virtual-organisations).
93+
94+
## Infrastructure status
95+
96+
[Check the system status](https://www.ugent.be/hpc/en/infrastructure/status)

0 commit comments

Comments
 (0)