Skip to content

Commit 10e7b74

Browse files
Phlip79claude
andauthored
cleanup: remove unused scatter_gather_tensors_in_pipeline argument (#4140)
Co-authored-by: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
1 parent 499266a commit 10e7b74

File tree

3 files changed

+2
-19
lines changed

3 files changed

+2
-19
lines changed

examples/academic_paper_scripts/sc21/run_figure_18.sh

Lines changed: 2 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -4,25 +4,12 @@
44
# Choose the case to run.
55
# ================================
66

7-
# Scatter-gather communication optimization options = [YES, NO].
8-
SCATTER_GATHER=YES
9-
107
# Batch size (global batch size) options = [12, 24, 36, ..., 60].
118
GBS=12
129

1310

1411

15-
16-
17-
# Set scatter-gather communication optimization options.
18-
if [ ${SCATTER_GATHER} == "YES" ]; then
19-
MEGATRON_EXTRA_PARAMS="--activations-checkpoint-method uniform --num-layers-per-virtual-pipeline-stage 2 "
20-
elif [ ${SCATTER_GATHER} == "NO" ]; then
21-
MEGATRON_EXTRA_PARAMS="--activations-checkpoint-method uniform --num-layers-per-virtual-pipeline-stage 2 --no-scatter-gather-tensors-in-pipeline "
22-
else
23-
echo "Invalid configuration"
24-
exit 1
25-
fi
12+
MEGATRON_EXTRA_PARAMS="--activations-checkpoint-method uniform --num-layers-per-virtual-pipeline-stage 2 "
2613

2714

2815
# Other params.
@@ -37,7 +24,7 @@ NNODES=12
3724

3825

3926
# Name of the job.
40-
export JOB_NAME=results_figure_18_scatter_gather_${SCATTER_GATHER}_batch_size_${GBS}
27+
export JOB_NAME=results_figure_18_batch_size_${GBS}
4128

4229

4330
# Import the configs.

examples/gpt3/gpt_config.yaml

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -218,7 +218,6 @@ overlap_grad_reduce: False
218218
align_grad_reduce: True
219219
overlap_param_gather: False
220220
align_param_gather: False
221-
scatter_gather_tensors_in_pipeline: True
222221
local_rank: null
223222
lazy_mpu_init: null
224223
empty_unused_memory_level: 0

megatron/training/arguments.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2696,9 +2696,6 @@ def _add_distributed_args(parser):
26962696
help='If not set, all PP stages will launch param all-gathers simultaneously. '
26972697
'Otherwise, each PP stage will independently launch as needed.',
26982698
dest='align_param_gather')
2699-
group.add_argument('--no-scatter-gather-tensors-in-pipeline', action='store_false',
2700-
help='If not set, use scatter/gather to optimize communication of tensors in pipeline.',
2701-
dest='scatter_gather_tensors_in_pipeline')
27022699
group.add_argument('--use-distributed-optimizer', action='store_true',
27032700
help='Use distributed optimizer.')
27042701
group.add_argument('--use-nccl-ub', action='store_true', dest='nccl_ub',

0 commit comments

Comments
 (0)