Skip to content

Commit daa574b

Browse files
committed
Modifications for the Istanbul workshop
1 parent ec0959a commit daa574b

File tree

1 file changed

+101
-48
lines changed
  • education/HADDOCK3/HADDOCK3-antibody-antigen

1 file changed

+101
-48
lines changed

education/HADDOCK3/HADDOCK3-antibody-antigen/index.md

Lines changed: 101 additions & 48 deletions
Original file line numberDiff line numberDiff line change
@@ -189,7 +189,7 @@ unzip HADDOCK3-antibody-antigen.zip
189189
</a>
190190

191191

192-
Unziping the file will create the `HADDOCK3-antibody-antigen-BioExcelSS2024` directory which should contain the following directories and files:
192+
Unziping the file will create the `HADDOCK3-antibody-antigen` directory which should contain the following directories and files:
193193

194194
* `pdbs`: a directory containing the pre-processed PDB files
195195
* `restraints`: a directory containing the interface information and the corresponding restraint files for HADDOCK3
@@ -208,35 +208,31 @@ This tutorial was last tested using HADDOCK3 version 2024.10.0b7. The provided p
208208

209209
<hr>
210210

211-
### ASEAN 2025 HPC school
211+
### EuroCC HPC workshop, Istanbul April 2025
212212

213-
We will be making use of the Fugaku supercomputer for this tutorial.
214-
The software and data required for this tutorial have been pre-installed on Fugaku.
215-
Please connect to Fugaku using your credentials either via ssh connection or from a web browser using OnDemand:
213+
We will be making use of the TRUBA computational resources for this tutorial.
214+
The software and data required for this tutorial have been pre-installed.
215+
Please connect to the system using your credentials either via ssh connection or from a web browser using OnDemand:
216216

217-
[https://ondemand.fugaku.r-ccs.riken.jp/](https://ondemand.fugaku.r-ccs.riken.jp/){:target="_blank"}
217+
[TRUBA onDemand interface](https://172.16.6.20/){:target="_blank"}
218218

219-
If using OnDemand, open then a terminal session.
219+
If using OnDemand, open then a terminal session using the `_arf Shell Access` menu.
220220

221-
In order to run the tutorial, go into you data directory, then copy and unzip the required data:
221+
In order to run the tutorial, go into you scratch directory, then unzip the required data:
222222

223223
<a class="prompt prompt-cmd">
224-
unzip /vol0601/data/hp240465/Materials/Life_Science/20250312_Bonvin/HADDOCK3-antibody-antigen.zip
224+
cd /arf/scratch/\<my\-username\><br>
225+
unzip ~egitin/HADDOCK/HADDOCK3-antibody-antigen.zip<br>
226+
cd HADDOCK3-antibody-antigen
225227
</a>
226228

227229
This will create the `HADDOCK3-antibody-antigen` directory with all necessary data and scripts and job examples ready for submission to the batch system.
228230

229-
HADDOCK3 has been pre-installed on the compute nodes. To test the installation, first create an interactive session on a node with:
231+
HADDOCK3 has been pre-installed. To activate the HADDOCK3 environment type:
230232

231233

232234
<a class="prompt prompt-cmd">
233-
pjsub \-\-interact \-L \"node=1\" \-L \"rscgrp=int\" \-L \"elapse=2:00:00\" \-\-sparam \"wait-time=600\" \-g hp240465 \-x PJM_LLIO_GFSCACHE=/vol0006:/vol0004
234-
</a>
235-
236-
Once the session is active, activate HADDOCK3 with:
237-
238-
<a class="prompt prompt-cmd">
239-
source /vol0601/data/hp240465/Materials/Life_Science/20250312_Bonvin/haddock3/.venv/bin/activate<br>
235+
source ~egitim\/HADDOCK\/haddock3\/.venv\/bin\/activate
240236
</a>
241237

242238
You can then test that `haddock3` is indeed accessible with:
@@ -838,7 +834,7 @@ The corresponding toml configuration file (provided in `workflows/docking-antibo
838834
# ====================================================================
839835

840836
# Directory in which the scoring will be done
841-
run_dir = "run1-CDR-NMR-CSP"
837+
run_dir = "run1"
842838

843839
# Compute mode
844840
mode = "local"
@@ -965,33 +961,81 @@ In in the first section of the workflow above we have a parameter `mode` definin
965961

966962
<hr>
967963

968-
#### Execution of HADDOCK3 on Fugaku (ASEAN HPC school)
964+
#### Execution of HADDOCK3 on the TRUBA resources (EuroCC Istanbul April 2024 workshop)
969965

970-
To execute the workflow on Fugaku, you can either start an interactive session or create a job file that will execute HADDOCK3 on a node, with HADDOCK3 running in local mode (the setup in the above configuration file with `mode="local"`) and harvesting all core of that node (`ncores=48`).
966+
To execute the HADDOCK3 workflow on the computational resources provided for this workshop,
967+
you should create an execution script contain specific requirements for the queueing system and the HADDOCK3 configuration and execution.
968+
Two scripts are provided with the data you unzipped, one for execution on the hamsri cluster and one for the barbun cluster:
971969

972-
**_Interactive session on a node:_**
970+
{% highlight shell %}
971+
run-haddock3-barbun.sh
972+
run-haddock3-hamsri.sh
973+
{% endhighlight %}
973974

974-
<a class="prompt prompt-cmd">
975-
pjsub --interact -L "node=1" -L "rscgrp=int" -L "elapse=2:00:00" --sparam "wait-time=600" -g hp240465 -x PJM_LLIO_GFSCACHE=/vol0006:/vol0004
976-
</a>
977975

978-
Once the session is active, activate HADDOCK3 with:
976+
Here is an example of such an execution script (also provided in the `HADDOCK3-antibody-antigen` directory as `run-haddock3-hamsri.sh`):
979977

980-
<a class="prompt prompt-cmd">
981-
source /vol0601/data/hp240465/Materials/Life_Science/20250312_Bonvin/haddock3/.venv/bin/activate<br>
982-
</a>
978+
{% highlight shell %}
979+
#!/bin/bash
980+
#SBATCH --nodes=1
981+
#SBATCH --tasks-per-node=54
982+
#SBATCH -C weka
983+
#SBATCH -p hamsi
984+
#SBATCH --time 00:30:00
983985

984-
You can then run the workflow with:
986+
source ~egitim/HADDOCK/haddock3/.venv/bin/activate
987+
haddock3 workflows/docking-antibody-antigen.cfg
988+
989+
{% endhighlight %}
990+
991+
This file should be submitted to the batch system using the `sbatch` command:
985992

986993
<a class="prompt prompt-cmd">
987-
haddock3 ./workflows/docking-antibody-antigen.cfg
994+
sbatch run-haddock3-hamsri.sh
988995
</a>
989996

997+
**_Note_** that batch submission is only possible from the `scratch` partition (`/arf/scratch/<my-home-directory>`)
990998

991-
**_Job submission to the batch system:_**
999+
And you can check the status in the queue using the `squeue`command.
9921000

1001+
This example run should take about 20 minutes to complete on a single node using 50 cores.
9931002

994-
For this execution mode you should create an execution script contain specific requirements for the queueing system and the HADDOCK3 configuration and execution. Here is an example of such an execution script (also provided in the `HADDOCK3-antibody-antigen` directory as `run-haddock3-fugaku.sh`):
1003+
1004+
<hr>
1005+
1006+
#### Execution of HADDOCK3 on Fugaku (ASEAN 2025 HPC school)
1007+
1008+
<details style="background-color:#DAE4E7">
1009+
<summary style="bold">
1010+
<b><i>Execution instructions for running HADDOCK3 on Fugaku</i></b> <i class="material-icons">expand_more</i>
1011+
</summary>
1012+
1013+
To execute the workflow on Fugaku, you can either start an interactive session or create a job file that will execute HADDOCK3 on a node,
1014+
with HADDOCK3 running in local mode (the setup in the above configuration file with <i>mode="local"</i>) and harvesting all core of that node (<i>ncores=48</i>).
1015+
<br>
1016+
<br>
1017+
<b>Interactive session on a node:</b>
1018+
<br>
1019+
{% highlight shell %}
1020+
pjsub --interact -L "node=1" -L "rscgrp=int" -L "elapse=2:00:00" --sparam "wait-time=600" -g hp240465 -x PJM_LLIO_GFSCACHE=/vol0006:/vol0004
1021+
{% endhighlight %}
1022+
1023+
Once the session is active, activate HADDOCK3 with:
1024+
1025+
{% highlight shell %}
1026+
source /vol0601/data/hp240465/Materials/Life_Science/20250312_Bonvin/haddock3/.venv/bin/activate<br>
1027+
{% endhighlight %}
1028+
1029+
You can then run the workflow with:
1030+
1031+
{% highlight shell %}
1032+
haddock3 ./workflows/docking-antibody-antigen.cfg
1033+
{% endhighlight %}
1034+
<b>Job submission to the batch system:</b>
1035+
<br>
1036+
<br>
1037+
For this execution mode you should create an execution script contain specific requirements for the queueing system and the HADDOCK3 configuration and execution.
1038+
Here is an example of such an execution script (also provided in the `HADDOCK3-antibody-antigen` directory as `run-haddock3-fugaku.sh`):
9951039

9961040
{% highlight shell %}
9971041
#!/bin/sh -x
@@ -1009,16 +1053,20 @@ haddock3 ./workflows/docking-antibody-antigen.cfg
10091053

10101054
This file should be submitted to the batch system using the `pjsub` command:
10111055

1012-
<a class="prompt prompt-cmd">
1056+
{% highlight shell %}
10131057
pjsub run-haddock3-fugaku.sh
1014-
</a>
1058+
{% endhighlight %}
1059+
10151060
<br>
10161061

10171062
And you can check the status in the queue using `pjstat`.
10181063

10191064
This run should take about 20 minutes to complete on a single node using 48 arm cores.
1065+
</details>
10201066

10211067

1068+
<hr>
1069+
10221070
#### Learn more about the various execution modes of haddock3
10231071

10241072
<hr>
@@ -1028,7 +1076,9 @@ This run should take about 20 minutes to complete on a single node using 48 arm
10281076
<b><i>Local execution</i></b> <i class="material-icons">expand_more</i>
10291077
</summary>
10301078

1031-
In this mode HADDOCK3 will run on the current system, using the defined number of cores (`ncores`) in the config file to a maximum of the total number of available cores on the system minus one. An example of the relevant parameters to be defined in the first section of the config file is:
1079+
In this mode HADDOCK3 will run on the current system, using the defined number of cores (<i>ncores</i>)
1080+
in the config file to a maximum of the total number of available cores on the system minus one.
1081+
An example of the relevant parameters to be defined in the first section of the config file is:
10321082

10331083
{% highlight toml %}
10341084
# compute mode
@@ -1039,21 +1089,21 @@ ncores = 50
10391089

10401090
In this mode HADDOCK3 can be started from the command line with as argument the configuration file of the defined workflow.
10411091

1042-
<a class="prompt prompt-cmd">
1043-
haddock3 \<my-workflow-configuration-file\>
1044-
</a>
1092+
{% highlight shell %}
1093+
haddock3 <my-workflow-configuration-file>
1094+
{% endhighlight %}
10451095

10461096
Alternatively redirect the output to a log file and send haddock3 to the background.
10471097

10481098

10491099
As an indication, running locally on an Apple M2 laptop using 10 cores, this workflow completed in 7 minutes.
10501100

10511101

1052-
<a class="prompt prompt-cmd">
1053-
haddock3 \<my-workflow-configuration-file\> \> haddock3.log &
1054-
</a>
1102+
{% highlight shell %}
1103+
haddock3 <my-workflow-configuration-file> > haddock3.log &
1104+
{% endhighlight %}
10551105

1056-
_**Note**_: This is also the execution mode that should be used for example when submitting the HADDOCK3 job to a node of a cluster, requesting X number of cores.
1106+
<b>Note</b>: This is also the execution mode that should be used for example when submitting the HADDOCK3 job to a node of a cluster, requesting X number of cores.
10571107

10581108
</details>
10591109

@@ -1081,15 +1131,18 @@ _**Note**_: This is also the execution mode that should be used for example when
10811131
cd $HOME/HADDOCK3-antibody-antigen
10821132

10831133
# execute
1084-
haddock3 \<my-workflow-configuration-file\>
1134+
haddock3 <my-workflow-configuration-file>
10851135
{% endhighlight %}
10861136
<br>
10871137

10881138

1089-
In this mode HADDOCK3 will typically be started on your local server (e.g. the login node) and will dispatch jobs to the batch system of your cluster. Two batch systems are currently supported: `slurm` and `torque` (defined by the `batch_type` parameter). In the configuration file you will
1090-
have to define the `queue` name and the maximum number of concurrent jobs sent to the queue (`queue_limit`).
1139+
In this mode HADDOCK3 will typically be started on your local server (e.g. the login node) and will dispatch jobs to the batch system of your cluster.
1140+
Two batch systems are currently supported: <i>slurm</i> and <i>torque</i> (defined by the <i>batch_type</i> parameter).
1141+
In the configuration file you will have to define the <i>queue</i> name and the maximum number of concurrent jobs sent to the queue (<i>queue_limit</i>).
10911142

1092-
Since HADDOCK3 single model calculations are quite fast, it is recommended to calculate multiple models within one job submitted to the batch system. The number of model per job is defined by the `concat` parameter in the configuration file. You want to avoid sending thousands of very short jobs to the batch system if you want to remain friend with your system administrators...
1143+
Since HADDOCK3 single model calculations are quite fast, it is recommended to calculate multiple models within one job submitted to the batch system.
1144+
he number of model per job is defined by the <i>concat</i> parameter in the configuration file.
1145+
You want to avoid sending thousands of very short jobs to the batch system if you want to remain friend with your system administrators...
10931146

10941147
An example of the relevant parameters to be defined in the first section of the config file is:
10951148

@@ -1117,8 +1170,8 @@ have to define the `queue` name and the maximum number of concurrent jobs sent t
11171170
</summary>
11181171

11191172

1120-
HADDOCK3 supports a parallel pseudo-MPI implementation. For this to work, the `mpi4py` library must have been installed at installation time.
1121-
Refer to the [MPI-related instructions](https://www.bonvinlab.org/haddock3/tutorials/mpi.html){:target="_blank"}.
1173+
HADDOCK3 supports a parallel pseudo-MPI implementation. For this to work, the <i>mpi4py</i> library must have been installed at installation time.
1174+
Refer to the (<a href="https://www.bonvinlab.org/haddock3/tutorials/mpi.html" target=new>MPI-related instructions</a>).
11221175

11231176
The execution mode should be set to `mpi` and the total number of cores should match the requested resources when submitting to the batch system.
11241177

0 commit comments

Comments
 (0)