Skip to content

Commit 01c130b

Browse files
committed
Updated READMEs
1 parent 23ca56c commit 01c130b

File tree

4 files changed

+105
-81
lines changed

4 files changed

+105
-81
lines changed

workflows/nt3_mlrMBO/README.md

Lines changed: 38 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -10,9 +10,25 @@ the Swift script to launch a NT3 run, and to
1010
2. Pass the validation loss from a NT3 run back to the running mlrBMO algorithm
1111
via the swift script.
1212

13-
The workflow ultimately produces a `final_res.Rds` serialized R object that
14-
contains the final best parameter values and various metadata about the
15-
parameter evaluations.
13+
For each run of the workflow, the following are produced:
14+
15+
* `final_res.Rds` - a serialized R object that
16+
contains the final best parameter values and various metadata about the
17+
parameter evaluations. This file will be written to the experiment directory (see below).
18+
* `experiment_start.json` - a json file containing experiment (i.e. workflow)
19+
level data (e.g. the start time, mlrMBO parameters, etc.). This will be
20+
written to the experiment directory.
21+
* `experiment_end.json` - a json file containing experiment (i.e. workflow)
22+
level data (e.g. the stop time, status info, etc). This will be
23+
written to the experiment directory.
24+
25+
For each run of the benchmark model, the following is produced:
26+
27+
* `run.json` - a json file containing data describing the individual run: the
28+
parameters for that run and per epoch details such as the validation loss. This
29+
file will be written to the output directory for that particular run (e.g.)
30+
`nt3_mlrMBO/experiments/E1/run_1_1_0/output/run.json`.
31+
1632

1733
## User requirements ##
1834

@@ -34,8 +50,8 @@ What you need to install to run the workflow:
3450
```
3551
mkdir -p Data/Pilot1
3652
cd Data/Pilot1
37-
wget ftp://ftp.mcs.anl.gov/pub/candle/public/benchmarks/Pilot1/normal-tumor/nt_train2.csv .
38-
wget ftp://ftp.mcs.anl.gov/pub/candle/public/benchmarks/Pilot1/normal-tumor/nt_test2.csv .
53+
wget ftp://ftp.mcs.anl.gov/pub/candle/public/benchmarks/Pilot1/normal-tumor/nt_train2.csv
54+
wget ftp://ftp.mcs.anl.gov/pub/candle/public/benchmarks/Pilot1/normal-tumor/nt_test2.csv
3955
```
4056

4157
## System requirements ##
@@ -146,18 +162,28 @@ the workflow is run, by defining which swift is actually run.
146162
* Set to `$EMEWS_PROJECT_ROOT\swift\ai_workflow3.swift` to run the benchmarks via a swift
147163
app function.
148164

149-
If you need to run the _ai_-version of the workflow, there is an addtional shell
150-
variable to set:
151-
152-
* `SCRIPT_FILE` - the path to the bash script that is used to launch the python
153-
benchmark runner code (e.g. `scripts/run_model.sh`).
154-
155165
If running on an HPC machine, set `PROCS`, `PPN`, `QUEUE`, `WALLTIME` and `MACHINE`
156166
as appropriate.
157167

158168
Lastly, see the TODOs in the launch script for any additional variables to set.
159169

160-
Running the *workflow script*:
170+
### App Invocation Shell Variables ###
171+
172+
If you need to run the _ai_-version of the workflow, there are two additional shell
173+
variables to set:
174+
175+
* `SWIFT_FILE` - the swift workflow file to run - set to
176+
`$EMEWS_PROJECT_ROOT\swift\ai_workflow3.swift` to run the benchmarks via a swift
177+
app function.
178+
* `SCRIPT_FILE` - the path to the bash script that is used to launch the python
179+
benchmark runner code (e.g. `scripts/run_model.sh`).
180+
* `LOG_SCRIPT_FILE` - the path to the bash script that is used to launch the python
181+
logging code code. By default these scripts are in the
182+
`Supervisor/workflows/common/sh` directory as they can be shared among the
183+
different workflows. See for example `Supervisor/workflows/common/sh/run_logger.sh`
184+
185+
186+
### Running the *workflow script* ###
161187

162188
The workflow is executed by running the launch script and passing it an
163189
'experiment id', i.e., `swift/workflow.sh <EXPID>` where `EXPID` is the

workflows/p2b1_mlrMBO/README.adoc

Lines changed: 0 additions & 37 deletions
This file was deleted.

workflows/p2b1_mlrMBO/README.md

Lines changed: 38 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -10,9 +10,24 @@ the Swift script to launch a P2B1 run, and to
1010
2. Pass the return value (TODO specify this) from a P2B1 run back to the running mlrBMO algorithm
1111
via the swift script.
1212

13-
The workflow ultimately produces a `final_res.Rds` serialized R object that
14-
contains the final best parameter values and various metadata about the
15-
parameter evaluations.
13+
For each run of the workflow, the following are produced:
14+
15+
* `final_res.Rds` - a serialized R object that
16+
contains the final best parameter values and various metadata about the
17+
parameter evaluations. This file will be written to the experiment directory (see below).
18+
* `experiment_start.json` - a json file containing experiment (i.e. workflow)
19+
level data (e.g. the start time, mlrMBO parameters, etc.). This will be
20+
written to the experiment directory.
21+
* `experiment_end.json` - a json file containing experiment (i.e. workflow)
22+
level data (e.g. the stop time, status info, etc). This will be
23+
written to the experiment directory.
24+
25+
For each run of the benchmark model, the following is produced:
26+
27+
* `run.json` - a json file containing data describing the individual run: the
28+
parameters for that run and per epoch details such as the validation loss. This
29+
file will be written to the output directory for that particular run (e.g.)
30+
`p2b1_mlrMBO/experiments/E1/run_1_1_0/output/run.json`.
1631

1732
## User requirements ##
1833

@@ -112,16 +127,16 @@ p2b1_mlrMBO/
112127
* `R/mlrMBO3.R` - the mlrMBO R code
113128
* `R/mlrMBO_utils.R` - utility functions used by the mlrMBO R code
114129
* `python/p2b1_runner.py` - python code called by the swift script to run P3B1.
115-
* `python/test/test.py` - python code for testing the p3b1_runner.
130+
* `python/test/test.py` - python code for testing the p2b1_runner.
116131
* `swift/workflow3.swift` - the swift workflow script
117132
* `swift/workflow.sh` - generic launch script to set the appropriate enviroment variables etc. and then launch the swift workflow script
118133
* `swift/cori_workflow3.sh` - launch script customized for the Cori supercomputer
119134
* `swift/cori_settings.sh` - settings for running on the Cori supercomputer
120135
* `swift/ai_workflow.sh` - launch script for running the app invocation ("ai") workflow (see below).
121136
* `swift/ai_workflow3.swift` - app invocation version (see below) of the swift workflow
122137
* `swift/theta_workflow.sh` - launch script for running on theta. This uses the app invocation workflow.
123-
* `scripts/theta_run_model.sh` - theta-specific bash script used to launch nt3_runner.py
124-
* `scripts/run_model.sh` - generic bash script used to to launch nt3_runner.py
138+
* `scripts/theta_run_model.sh` - theta-specific bash script used to launch p2b1_runner.py
139+
* `scripts/run_model.sh` - generic bash script used to to launch p2b1_runner.py
125140

126141
## Running the Workflow ##
127142

@@ -160,32 +175,30 @@ int design_size = toint(argv("ds", "10"));
160175
string param_set = argv("param_set_file");
161176
```
162177

163-
In addition, the following variable is used to determine which version of
164-
the workflow is run, by defining which swift is actually run.
165-
166-
* `SWIFT_FILE` - the swift workflow file to run
167-
* Set to `$EMEWS_PROJECT_ROOT\swift\workflow3.swift` to run the benchmarks via swift's integrated python.
168-
* Set to `$EMEWS_PROJECT_ROOT\swift\ai_workflow3.swift` to run the benchmarks via a swift
169-
app function.
170-
171-
If you need to run the _ai_-version of the workflow, there is an addtional shell
172-
variable to set:
173-
174-
* `SCRIPT_FILE` - the path to the bash script that is used to launch the python
175-
benchmark runner code (e.g. `scripts/run_model.sh`).
176-
177178
If running on an HPC machine, set `PROCS`, `PPN`, `QUEUE`, `WALLTIME` and `MACHINE`
178179
as appropriate.
179180

180181
Lastly, see the TODOs in the launch script for any additional variables to set.
181182

182-
Running the *workflow script*:
183+
### App Invocation Shell Variables ###
184+
185+
If you need to run the _ai_-version of the workflow, there are two additional shell
186+
variables to set:
187+
188+
* `SCRIPT_FILE` - the path to the bash script that is used to launch the python
189+
benchmark runner code (e.g. `scripts/run_model.sh`).
190+
* `LOG_SCRIPT_FILE` - the path to the bash script that is used to launch the python
191+
logging code code. By default these scripts are in the
192+
`Supervisor/workflows/common/sh` directory as they can be shared among the
193+
different workflows. See for example `Supervisor/workflows/common/sh/run_logger.sh`
194+
195+
### Running the *workflow script* ###
183196

184197
The workflow is executed by running the launch script and passing it an
185198
'experiment id', i.e., `swift/workflow.sh <EXPID>` where `EXPID` is the
186199
experiment ID (provide any token you want). The workflow
187200
output, various swift related files, and the `final_res.Rds` file will be written
188-
into a `nt3_mlrMBO/experiments/X` directory where X is the experiment id. A copy
201+
into a `p2b1_mlrMBO/experiments/X` directory where X is the experiment id. A copy
189202
of the launch script that was used to launch the workflow will also be written
190203
to this directory.
191204

@@ -264,7 +277,7 @@ In addition to final_res.Rds, for each run the workflow writes out the hyperpara
264277
used in that run to a `parameters.txt` file in each run's instance directory.
265278
`parameters.txt` can be used to run the model outside of the workflow using
266279
the `--config_file` command line argument. For example,
267-
`python nt3_baseline_keras2.py --config_file parameters.txt`
280+
`python p2b1_baseline_keras2.py --config_file parameters.txt`
268281

269282
### Running on Cori ###
270283

@@ -276,7 +289,7 @@ for the EQ/R swift extension.
276289

277290
* Compile the EQ/R swift-t extension.
278291
```
279-
cd Supervisor/workflows/nt3_mlrMBO/ext/EQ-R/eqr
292+
cd Supervisor/workflows/p2b1_mlrMBO/ext/EQ-R/eqr
280293
./cori_build.sh
281294
```
282295

@@ -292,7 +305,7 @@ debugging runs and will need to be changed for a production run.
292305
An example:
293306

294307
```
295-
cd Supervisor/workflows/nt3_mlrMBO/swift
308+
cd Supervisor/workflows/p2b1_mlrMBO/swift
296309
source cori_settings.sh
297310
./cori_workflow.sh T1
298311
```

workflows/p3b1_mlrMBO/README.md

Lines changed: 29 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -10,9 +10,25 @@ the Swift script to launch a P3B1 run, and to
1010
2. Pass the validation loss from a P3B1 run back to the running mlrBMO algorithm
1111
via the swift script.
1212

13-
The workflow ultimately produces a `final_res.Rds` serialized R object that
13+
For each run of the workflow, the following are produced:
14+
15+
* `final_res.Rds` - a serialized R object that
1416
contains the final best parameter values and various metadata about the
15-
parameter evaluations.
17+
parameter evaluations. This file will be written to the experiment directory (see below).
18+
* `experiment_start.json` - a json file containing experiment (i.e. workflow)
19+
level data (e.g. the start time, mlrMBO parameters, etc.). This will be
20+
written to the experiment directory.
21+
* `experiment_end.json` - a json file containing experiment (i.e. workflow)
22+
level data (e.g. the stop time, status info, etc). This will be
23+
written to the experiment directory.
24+
25+
For each run of the benchmark model, the following is produced:
26+
27+
* `run.json` - a json file containing data describing the individual run: the
28+
parameters for that run and per epoch details such as the validation loss. This
29+
file will be written to the output directory for that particular run (e.g.)
30+
`p3b1_mlrMBO/experiments/E1/run_1_1_0/output/run.json`.
31+
1632

1733
## User requirements ##
1834

@@ -31,7 +47,7 @@ within `X/Benchmarks`
3147
```
3248
mkdir -p Data/P3B1
3349
cd Data/P3B1
34-
wget http://ftp.mcs.anl.gov/pub/candle/public/benchmarks/P3B1/P3B1_data.tgz .
50+
wget ftp://ftp.mcs.anl.gov/pub/candle/public/benchmarks/P3B1/P3B1_data.tgz
3551
tar -xf P3B1_data.tgz
3652
```
3753

@@ -140,19 +156,25 @@ as appropriate.
140156

141157
Lastly, see the TODOs in the launch script for any additional variables to set.
142158

143-
If you need to run the _ai_-version of the workflow, there is an addtional shell
144-
variable to set:
159+
### App Invocation Shell Variables ###
160+
161+
If you need to run the _ai_-version of the workflow, there are two additional shell
162+
variables to set:
145163

146164
* `SCRIPT_FILE` - the path to the bash script that is used to launch the python
147165
benchmark runner code (e.g. `scripts/run_model.sh`).
166+
* `LOG_SCRIPT_FILE` - the path to the bash script that is used to launch the python
167+
logging code code. By default these scripts are in the
168+
`Supervisor/workflows/common/sh` directory as they can be shared among the
169+
different workflows. See for example `Supervisor/workflows/common/sh/run_logger.sh`
148170

149-
Running the *workflow script*:
171+
### Running the *workflow script* ###
150172

151173
The workflow is executed by running the launch script and passing it an
152174
'experiment id', i.e., `swift/workflow.sh <EXPID>` where `EXPID` is the
153175
experiment ID (provide any token you want). The workflow
154176
output, various swift related files, and the `final_res.Rds` file will be written
155-
into a `P3B1_mlrMBO/experiments/X` directory where X is the experiment id. A copy
177+
into a `p3b1_mlrMBO/experiments/X` directory where X is the experiment id. A copy
156178
of the launch script that was used to launch the workflow will also be written
157179
to this directory.
158180

0 commit comments

Comments
 (0)