|
| 1 | +# Infrastructure |
| 2 | + |
| 3 | +## Tier2 clusters of Ghent University |
| 4 | + |
| 5 | +The Stevin computing infrastructure consists of several Tier2 clusters |
| 6 | +which are hosted in the S10 datacenter of Ghent University. |
| 7 | + |
| 8 | +This infrastructure is co-financed by FWO and Department of Economy, |
| 9 | +Science and Innovation (EWI). |
| 10 | + |
| 11 | +## Tier-2 login nodes |
| 12 | + |
| 13 | +Log in to the HPC-UGent Tier-2 infrastructure via [https://login.hpc.ugent.be](https://login.hpc.ugent.be) |
| 14 | +or using SSH via `login.hpc.ugent.be`. |
| 15 | + |
| 16 | +more info on using the web portal you can find [here](web_portal), |
| 17 | +and about connection with SSH [here](connecting). |
| 18 | + |
| 19 | +## Tier-2 compute clusters |
| 20 | + |
| 21 | +### CPU clusters |
| 22 | + |
| 23 | +The HPC-UGent Tier-2 infrastructure currently included several standard |
| 24 | +CPU-only clusters, of different generations (listed from old to new). |
| 25 | + |
| 26 | +For basic information on using these clusters, see our |
| 27 | +[documentation](running_batch_jobs.md). |
| 28 | + |
| 29 | +| ***cluster name*** | ***# nodes*** | ***Processor architecture*** | ***Usable memory/node*** | ***Local diskspace/node*** | ***Interconnect*** | ***Operating system*** | |
| 30 | +| --- | --- | --- | --- | --- | --- | --- | |
| 31 | +| skitty | 72 | 2 x 18-core Intel Xeon Gold 6140 (Skylake @ 2.3 GHz) | 177 GiB | 1 TB & 240 GB SSD | EDR InfiniBand | RHEL 9 | |
| 32 | +| doduo (default cluster) | 128 | 2x 48-core AMD EPYC 7552 (Rome @ 2.2 GHz) | 250 GiB | 180GB SSD | HDR-100 InfiniBand | RHEL 8 | |
| 33 | +| gallade (*) | 16 | 2x 64-core AMD EPYC 7773X (Milan-X @ 2.2 GHz) | 940 GiB | 1.5 TB NVME | HDR-100 InfiniBand | RHEL 9 | |
| 34 | +| shinx | 48 | 2x 96-core AMD EPYC 9654 (Genoa @ 2.4 GHz) | 370 GiB | 500GB NVME | NDR-200 InfiniBand | RHEL 9 | |
| 35 | + |
| 36 | +(*) also see this [extra information](./only/gent/2023/donphan-gallade#gallade-large-memory-cluster) |
| 37 | + |
| 38 | +### Interactive debug cluster |
| 39 | + |
| 40 | + |
| 41 | +A special-purpose interactive debug cluster is available, |
| 42 | +where you should always be able to get a job running quickly, |
| 43 | +**without waiting in the queue**. |
| 44 | + |
| 45 | +Intended usage is mainly for interactive work, |
| 46 | +either via an interactive job or using the [HPC-UGent web portal](web_portal). |
| 47 | + |
| 48 | +This cluster is heavily over-provisioned, so jobs may |
| 49 | +run slower if the cluster is used more heavily. |
| 50 | + |
| 51 | +Strict limits are in place per user: |
| 52 | + * max. 5 jobs in queue |
| 53 | + * max. 3 jobs running |
| 54 | + * max. of 8 cores and 27GB of memory in total for running jobs |
| 55 | + |
| 56 | +For more information, see our [documentation](interactive_gent). |
| 57 | + |
| 58 | +| ***cluster name*** | ***# nodes*** | ***Processor architecture*** | ***Usable memory/node*** | ***Local diskspace/node*** | ***Interconnect*** | ***Operating system*** | |
| 59 | +| --- | --- | --- | --- | --- | --- | --- | |
| 60 | +| donphan (*) | 16 | 2 x 18-core Intel Xeon Gold 6240 (Cascade Lake @ 2.6 GHz) + 1x shared NVIDIA Ampere A2 GPU (16GB GPU memory) | 738 GiB | 1.6 TB NVME | HDR-100 Infiniband | RHEL 8 | |
| 61 | + |
| 62 | +(*) also see this [extra information](./only/gent/2023/donphan-gallade#donphan-debuginteractive-cluster) |
| 63 | + |
| 64 | +### GPU clusters |
| 65 | + |
| 66 | +GPU clusters are available in the HPC-UGent Tier-2 infrastructure, |
| 67 | +with different generations of NVIDIA GPUs. |
| 68 | + |
| 69 | +These are well suited for specific workloads, with software that |
| 70 | +can leverage the GPU resources (like TensorFlow, PyTorch, GROMACS, AlphaFold, etc.). |
| 71 | + |
| 72 | +For more information on using these clusters, see our documentation. |
| 73 | + |
| 74 | +| ***cluster name*** | ***# nodes*** | ***Processor architecture & GPUs*** | ***Usable memory/node*** | ***Local diskspace/node*** | ***Interconnect*** | ***Operating system*** | |
| 75 | +| --- | --- | --- | --- | --- | --- | --- | |
| 76 | +| joltik | 10 | 2x 16-core Intel Xeon Gold 6242 (Cascade Lake @ 2.8 GHz) + 4x NVIDIA Volta V100 GPUs (32GB GPU memory) | 256 GiB | 800GB SSD | double EDR Infiniband | RHEL 9 | |
| 77 | +| accelgor | 9 | 2x 24-core AMD EPYC 7413 (Milan @ 2.2 GHz) + 4x NVIDIA Ampere A100 GPUs (80GB GPU memory) | 500 GiB | 180GB SSD | HDR InfiniBand | RHEL 8 | |
| 78 | + |
| 79 | + |
| 80 | +## Tier-2 shared storage |
| 81 | + |
| 82 | +| ***Filesystem name*** | ***Intended usage*** | ***Total storage space*** | ***Personal storage space*** | ***VO storage space (^)*** | |
| 83 | +| ---| --- |---| --- | --- | |
| 84 | +| $VSC_HOME | Home directory, entry point to the system | 90 TB | 3GB (fixed) | (none) | |
| 85 | +| $VSC_DATA | Long-term storage of large data files | 1.9 PB | 25GB (fixed) | 250GB | |
| 86 | +| $VSC_SCRATCH | Temporary fast storage of 'live' data for calculations | 1.7 PB | 25GB (fixed) | 250GB | |
| 87 | +| $VSC_SCRATCH_ARCANINE | Temporary very fast storage of 'live' data for calculations (recommended for very I/O-intensive jobs) | 70 TB NVME | (none) | upon request | |
| 88 | + |
| 89 | + |
| 90 | +^ Storage space for a group of users (Virtual Organisation or VO for short) can be |
| 91 | +increased significantly on request. For more information, see our |
| 92 | +[documentation](running_jobs_with_input_output_data#virtual-organisations). |
| 93 | + |
| 94 | +## Infrastructure status |
| 95 | + |
| 96 | +[Check the system status](https://www.ugent.be/hpc/en/infrastructure/status) |
0 commit comments