Skip to content

Commit d74071a

Browse files
committed
[fix] remove useless param in doc; fix gpt2 qkv test;
1 parent feca06e commit d74071a

File tree

3 files changed

+3
-5
lines changed

3 files changed

+3
-5
lines changed

docs/source/en/features/zerobubble_pipeline_parallelism.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,11 +36,10 @@ Build our model and Optimizer. We created a Llama with 8 Decoder-Layer. Then, in
3636
```python
3737
# Global Param
3838
NUM_BATCH = 8
39-
NUM_TOK_PER_BATCH, NUM_EXPERTS = 4, 4
39+
NUM_TOK_PER_BATCH = 4
4040
NUM_LAYERS = 8
4141
HIDDEN_SIZE_PER_HEAD = 4
4242
NUM_HEADS = 4
43-
TOP_K = 1
4443
# Init Llama from huggingface
4544
configuration = LlamaConfig(
4645
hidden_size=HIDDEN_SIZE_PER_HEAD * NUM_HEADS,

docs/source/zh-Hans/features/zerobubble_pipeline_parallelism.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,11 +37,10 @@ colossalai.launch(rank=rank, world_size=world_size, host="localhost", port=port,
3737
```python
3838
# Global Param
3939
NUM_BATCH = 8
40-
NUM_TOK_PER_BATCH, NUM_EXPERTS = 4, 4
40+
NUM_TOK_PER_BATCH = 4
4141
NUM_LAYERS = 8
4242
HIDDEN_SIZE_PER_HEAD = 4
4343
NUM_HEADS = 4
44-
TOP_K = 1
4544
# Init Llama from huggingface
4645
configuration = LlamaConfig(
4746
hidden_size=HIDDEN_SIZE_PER_HEAD * NUM_HEADS,

tests/test_shardformer/test_layer/test_gpt2_qkv_fused_linear_1d.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -188,7 +188,7 @@ def check_linear_conv_1d_with_weight_grad_store(lazy_init: bool, seq_parallel_mo
188188
assert_close(linear.weight.grad, linear_base.weight.grad)
189189

190190

191-
@parameterize("lazy_init", [False])
191+
@parameterize("lazy_init", [False, True])
192192
@parameterize("seq_parallel_mode", ["split_gather", None])
193193
def check_gpt2_qkv_fused_linear_1d(lazy_init: bool, seq_parallel_mode: bool):
194194
check_linear_conv_1d_col(lazy_init, seq_parallel_mode)

0 commit comments

Comments
 (0)