Skip to content

Commit 1e48900

Browse files
committed
Merge branch 'main' into release/3.3
2 parents 999bd79 + 5ff0f86 commit 1e48900

File tree

4 files changed

+2
-6
lines changed

4 files changed

+2
-6
lines changed

docs/source/GetStarted/SWIFT安装.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,6 @@
88
pip install 'ms-swift'
99
# 使用评测
1010
pip install 'ms-swift[eval]' -U
11-
# 使用序列并行
12-
pip install 'ms-swift[seq_parallel]' -U
1311
# 全能力
1412
pip install 'ms-swift[all]' -U
1513
```

docs/source/Instruction/预训练与微调.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -65,8 +65,8 @@ ms-swift使用了分层式的设计思想,用户可以使用命令行界面、
6565
- Any-to-Any模型训练:参考[这里](https://github.com/modelscope/swift/blob/main/examples/train/all_to_all)
6666
- 其他能力:
6767
- 数据流式读取: 在数据量较大时减少内存使用。参考[这里](https://github.com/modelscope/swift/blob/main/examples/train/streaming/train.sh)
68-
- 序列并行: 参考[这里](https://github.com/modelscope/swift/blob/main/examples/train/sequence_parallel)
6968
- packing: 将多个序列拼成一个,让每个训练样本尽可能接近max_length,提高显卡利用率,参考[这里](https://github.com/modelscope/swift/blob/main/examples/train/packing/train.sh)
69+
- 长文本训练: 参考[这里](https://github.com/modelscope/swift/blob/main/examples/train/long_text)
7070
- lazy tokenize: 在训练期间对数据进行tokenize而不是在训练前tokenize(多模态模型可以避免在训练前读入所有多模态资源),这可以避免预处理等待并节约内存。参考[这里](https://github.com/modelscope/swift/blob/main/examples/train/lazy_tokenize/train.sh)
7171

7272
小帖士:

docs/source_en/GetStarted/SWIFT-installation.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,6 @@ You can install it using pip:
88
pip install 'ms-swift'
99
# For evaluation usage
1010
pip install 'ms-swift[eval]' -U
11-
# For sequence parallel usage
12-
pip install 'ms-swift[seq_parallel]' -U
1311
# Full capabilities
1412
pip install 'ms-swift[all]' -U
1513
```

docs/source_en/Instruction/Pre-training-and-Fine-tuning.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -68,8 +68,8 @@ Additionally, we offer a series of scripts to help you understand the training c
6868
- Any-to-Any Model Training: Refer to [here](https://github.com/modelscope/swift/blob/main/examples/train/all_to_all).
6969
- Other Capabilities:
7070
- Streaming Data Reading: Reduces memory usage when handling large datasets. Refer to [here](https://github.com/modelscope/swift/blob/main/examples/train/streaming/train.sh).
71-
- Sequence Parallelism: Refer to [here](https://github.com/modelscope/swift/blob/main/examples/train/sequence_parallel).
7271
- Packing: Combines multiple sequences into one, making each training sample as close to max_length as possible to improve GPU utilization. Refer to [here](https://github.com/modelscope/swift/blob/main/examples/train/packing/train.sh).
72+
- Long Text Training: Refer to [here](https://github.com/modelscope/swift/blob/main/examples/train/long_text).
7373
- Lazy Tokenize: Performs tokenization during training instead of pre-training (for multi-modal models, this avoids the need to load all multi-modal resources before training), which can reduce preprocessing wait times and save memory. Refer to [here](https://github.com/modelscope/swift/blob/main/examples/train/lazy_tokenize/train.sh).
7474

7575

0 commit comments

Comments
 (0)