Skip to content

Commit 7ebfe18

Browse files
authored
Merge pull request #243 from AliceO2Group/jgrosseo-patch-1
Update hyperlooppolicy.md
2 parents 6f4d63b + 0200deb commit 7ebfe18

File tree

2 files changed

+19
-1
lines changed

2 files changed

+19
-1
lines changed

docs/hyperloop/hyperlooppolicy.md

Lines changed: 19 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ sort: 2
33
title: Hyperloop policy
44
---
55

6-
## <a name="hyperlooppolicy"></a>General resource usage
6+
## <a name="generalresourcerules"></a>General resource rules
77

88
The very large amount of data that will be collected in Run 3 represents a challenge for analysis, for both the CPU needs and the read data from storage, and therefore a resource usage policy has been put in place to ensure proper use of computing resources. The policy has been openly discussed in multiple meetings, including ALICE weeks, and is subject to adjustments as necessary and as the collaboration gains experience with the Run 3 analysis.
99

@@ -19,3 +19,21 @@ In general, four categories of trains exist:
1919
* Trains that are lower than 1.5y in CPU usage and loop over less than 200 TB are free to execute and can be executed on Hyperloop via autosubmission. In a certain region between 30-200 TB, slightly more than 1.5y in CPU time is allowed as long as performance is better than 1 MB/s (green shaded area).
2020
* Trains that loop over more than 200 TB and have a throughput of at least 3 MB/s can run, provided that PWG conveners approve them (pink shaded area).
2121
* Heavy trains looping over datasets bigger than 200 TB or with low throughput above 30 TB, as marked in the blue region in the plot, require Physics Board approval to run. For those trains, a special analysis budget can be negotiated with the Physics Board.
22+
23+
## <a name="implementation"></a>Implementation in Hyperloop datasets
24+
25+
In practice the chart above is mapped on a number of distinct resource groups which determine the limits assigned to each dataset:
26+
27+
<div align="center">
28+
<img src="../images/resourcetable.png" width="80%">
29+
</div>
30+
31+
The smaller the dataset size, the more often it is automatically submitted per week and the more often you are allowed to run on it per week.
32+
33+
## <a name="deriveddata"></a>Derived data
34+
35+
Derived datasets can be created on Hyperloop which are by construction much smaller than the original datasets. Those are advantagous because steps which are identical in each analysis train run (e.g. event selection and centrality calculation, secondary-vertex finding) are only executed once which saves CPU. Furthermore, as the size is smaller such trains cause less load on the storages.
36+
37+
As an example, you can imagine that you run a derived data train on a dataset of 500 TB where you need explicit approval. Say you have a reduction factor of 100, then your output derived data is about 5 TB. You will be allowed to run on that dataset much more frequent, see the table above.
38+
39+
However, it may be that your analysis remains CPU hungry even if the input dataset is reduced. By default the operators who create your dataset will assign the resources based on the table above. The person who has requested the dataset can request to change that assignement. This will increase the CPU limit but reduce the trains/week and auto-submission slots by moving down one row in the table above. Obviously, running on a derived dataset should never take more resources than the limits that were their on the original parent dataset.

docs/images/resourcetable.png

28.9 KB
Loading

0 commit comments

Comments
 (0)