Skip to content

Commit 602714e

Browse files
committed
better naming
1 parent a1bdd3f commit 602714e

File tree

1 file changed

+12
-12
lines changed

1 file changed

+12
-12
lines changed

tools/emr/submit_datagen_job.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@
2424
'use_spot': True,
2525
'master_instance_type': 'r6gd.2xlarge',
2626
'instance_type': 'r6gd.4xlarge',
27-
'sf_per_executors': 3e3,
28-
'sf_per_partitions': 1e2,
27+
'sf_per_executor': 3e3,
28+
'sf_per_partition': 1e2,
2929
'az': 'us-west-2c',
3030
'yes': False,
3131
'ec2_key': None,
@@ -68,9 +68,9 @@ def submit_datagen_job(name,
6868
use_spot,
6969
instance_type,
7070
executors,
71-
sf_per_executors,
71+
sf_per_executor,
7272
partitions,
73-
sf_per_partitions,
73+
sf_per_partition,
7474
master_instance_type,
7575
az,
7676
emr_release,
@@ -106,10 +106,10 @@ def submit_datagen_job(name,
106106
}
107107

108108
if executors is None:
109-
executors = max(min_num_workers, min(max_num_workers, ceil(sf / sf_per_executors)))
109+
executors = max(min_num_workers, min(max_num_workers, ceil(sf / sf_per_executor)))
110110

111111
if partitions is None:
112-
partitions = max(min_num_threads, ceil(sf / sf_per_partitions))
112+
partitions = max(min_num_threads, ceil(sf / sf_per_partition))
113113

114114
spark_defaults_config = {
115115
'spark.serializer': 'org.apache.spark.serializer.KryoSerializer',
@@ -265,20 +265,20 @@ def submit_datagen_job(name,
265265
type=int,
266266
help=f"Total number of Spark executors."
267267
)
268-
executor_args.add_argument("--sf-per-executors",
268+
executor_args.add_argument("--sf-per-executor",
269269
type=float,
270-
default=defaults['sf_per_executors'],
271-
help=f"Number of Spark executors per scale factor. Default: {defaults['sf_per_executors']}"
270+
default=defaults['sf_per_executor'],
271+
help=f"Number of scale factors per Spark executor. Default: {defaults['sf_per_executor']}"
272272
)
273273
partitioning_args = parser.add_mutually_exclusive_group()
274274
partitioning_args.add_argument("--partitions",
275275
type=int,
276276
help=f"Total number of Spark partitions to use when generating the dataset."
277277
)
278-
partitioning_args.add_argument("--sf-per-partitions",
278+
partitioning_args.add_argument("--sf-per-partition",
279279
type=float,
280-
default=defaults['sf_per_partitions'],
281-
help=f"Number of Spark partitions per scale factor to use when generating the dataset. Default: {defaults['sf_per_partitions']}"
280+
default=defaults['sf_per_partition'],
281+
help=f"Number of scale factors per Spark partitions. Default: {defaults['sf_per_partition']}"
282282
)
283283

284284
parser.add_argument('--', nargs='*', help='Arguments passed to LDBC SNB Datagen', dest="arg")

0 commit comments

Comments
 (0)