Skip to content

Commit 70a1ba6

Browse files
Add commands to configs and update readme (#290)
* Add commands to configs and update readme * update config
1 parent dc21a19 commit 70a1ba6

File tree

5 files changed

+9
-3
lines changed

5 files changed

+9
-3
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ The primary purpose of the Forge ecosystem is to delineate infra concerns from m
99
Key features:
1010
- Usability for rapid research (isolating the RL loop from infrastructure)
1111
- Hackability for power users (all parts of the RL loop can be easily modified without interacting with infrastructure)
12-
- Scalability (ability so shift between async and syncronous training and across thousands of GPUs)
12+
- Scalability (ability to shift between async and synchronous training and across thousands of GPUs)
1313

1414
> ⚠️ **Early Development Warning** Forge is currently in an experimental
1515
> stage. You should expect bugs, incomplete features, and APIs that may change
@@ -47,7 +47,7 @@ Optional: By default, the packages installation uses conda. If user wants to ins
4747
After install, you can run the following command and should see output confirming GRPO training is running (you need a minimum 3 GPU devices):
4848

4949
```
50-
python -m apps.grpo.main --config apps/grpo/qwen3_1_7b.yaml
50+
python -m apps.grpo.main --config apps/grpo/qwen3_1_7b.yaml
5151
```
5252

5353
If you need to re-build the wheels for whatever reason, you can do so with:

apps/grpo/qwen3_multinode.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
# GRPO Training Configuration
22
# Currently a fork of the main yaml, this just shows
33
# placement of trainer and inference servers on separate hosts.
4+
# >>> python -m apps.grpo.main --config apps/grpo/qwen3_multinode.yaml
45

56
# Global configuration
67
group_size: 8

apps/vllm/deepseek_r1.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,5 @@
1+
# >>> python -m apps.vllm.main --config apps/vllm/deepseek_r1.yaml
2+
13
# NOTE - this won't work until we have proper HostMesh support
24
policy:
35
engine_config:

apps/vllm/llama3_8b.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,5 @@
1+
# >>> python -m apps.vllm.main --config apps/vllm/llama3_8b.yaml
2+
13
policy:
24
engine_config:
35
model: "meta-llama/Llama-3.1-8B-Instruct"

apps/vllm/qwen2_5_32b.yaml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,5 @@
1+
# >>> python -m apps.vllm.main --config apps/vllm/qwen2_5_32b.yaml
2+
13
policy:
24
engine_config:
35
model: "Qwen/Qwen2.5-32B"
@@ -12,7 +14,6 @@ policy:
1214
services:
1315
policy:
1416
procs: 4
15-
hosts: 1
1617
num_replicas: 1
1718
with_gpus: true
1819

0 commit comments

Comments
 (0)