Skip to content

Commit 9bf5208

Browse files
authored
update docs (#1211)
1 parent 9303bfd commit 9bf5208

File tree

11 files changed

+960
-8
lines changed

11 files changed

+960
-8
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010
1. **High-Performance Training**: Supports efficient training in various modes by connecting Megatron with SGLang;
1111
2. **Flexible Data Generation**: Enables arbitrary training data generation workflows through custom data generation interfaces and server-based engines.
1212

13-
slime is the RL-framework behind [GLM-4.5](https://z.ai/blog/glm-4.5) and [GLM-4.6](https://z.ai/blog/glm-4.6) and apart from models from Z.ai, we also supports the following models:
13+
slime is the RL-framework behind [GLM-4.7](https://z.ai/blog/glm-4.7), [GLM-4.6](https://z.ai/blog/glm-4.6), [GLM-4.5](https://z.ai/blog/glm-4.5) and apart from models from Z.ai, we also supports the following models:
1414
- Qwen3 series (Qwen3Next, Qwen3MoE, Qwen3), Qwen2.5 series;
1515
- DeepSeek V3 series (DeepSeek V3, V3.1, DeepSeek R1);
1616
- Llama 3.

README_zh.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010
1. **高性能训练**:通过连接 Megatron 与 SGLang,支持各种模式的高效训练;
1111
2. **灵活的数据生成**:通过自定义数据生成接口以及 server based engine,实现任意的数据训练数据生成流程。
1212

13-
slime 是 [GLM-4.5](https://z.ai/blog/glm-4.5)[GLM-4.6](https://z.ai/blog/glm-4.6) 背后的 RL 训练框架,除此之外,slime 还支持:
13+
slime 是 [GLM-4.7](https://z.ai/blog/glm-4.7)[GLM-4.6](https://z.ai/blog/glm-4.6)[GLM-4.5](https://z.ai/blog/glm-4.5) 背后的 RL 训练框架,除此之外,slime 还支持:
1414
- Qwen3 系列 (Qwen3Next, Qwen3MoE, Qwen3), Qwen2.5 系列;
1515
- DeepSeek V3 系列 (DeepSeek V3, V3.1, DeepSeek R1);
1616
- Llama 3。
Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,5 @@
1+
# PD Disaggregation
2+
3+
Slime supports Prefill and Decode disaggregation (PD Disaggregation).
4+
5+
You can set the number of servers used for Prefill by setting the `--prefill-num-servers` argument.

docs/en/advanced/speculative-decoding.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ Speculative decoding is a key optimization for speeding up rollouts. Instead of
44

55
## Accelerating Inference with Speculative Decoding
66

7-
For models with MTP layers (e.g., GLM-4.6, DeepSeek-V3/R1), simply add:
7+
For models with MTP layers (e.g., GLM-4.7, DeepSeek-V3/R1), simply add:
88

99
```bash
1010
--sglang-speculative-algorithm EAGLE

0 commit comments

Comments
 (0)