You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
#### A PyTorch native agentic library for RL post-training and agentic development
4
+
#### A PyTorch native agentic library for RL post-training and agentic development that lets you focus on algorithms instead of writing infra code.
5
5
6
6
## Overview
7
-
Forge was built with one core principle in mind: researchers should write algorithms, not infrastructure. Forge introduces a “service”-centric architecture that provides the right abstractions for distributed complexity. When you need fine-grained control over placement, fault handling or communication patterns, the primitives are there. When you don’t, you can focus purely on your RL algorithm.
7
+
The primary purpose of the Forge ecosystem is to delineate infra concerns from model concerns thereby making RL experimentation easier. Forge delivers this by providing clear RL abstractions and one scalable implementation of these abstractions. When you need fine-grained control over placement, fault handling/redirecting training loads during a run, or communication patterns, the primitives are there. When you don’t, you can focus purely on your RL algorithm.
8
8
9
9
Key features:
10
10
- Usability for rapid research (isolating the RL loop from infrastructure)
@@ -18,15 +18,19 @@ Key features:
18
18
> work. It's recommended that you signal your intention to contribute in the
19
19
> issue tracker, either by filing a new issue or by claiming an existing one.
20
20
21
-
## 📖 Documentation
21
+
## 📖 Documentation (Coming Soon)
22
22
23
-
View Forge's hosted documentation [at this link](https://meta-pytorch.org/forge/).
23
+
View Forge's hosted documentation (coming soon)
24
+
25
+
## Tutorials
26
+
27
+
You can also find our notebook tutorials (coming soon)
24
28
25
29
## Installation
26
30
27
31
### Basic
28
32
29
-
Forge requires the latest PyTorch nightly with Monarch, vLLM, and torchtitan. For convenience,
33
+
Forge requires the latest PyTorch nightly with [Monarch](https://github.com/meta-pytorch/monarch), [vLLM](https://github.com/vllm-project/vllm), and [torchtitan](https://github.com/pytorch/torchtitan). For convenience,
30
34
we have pre-packaged these dependencies as wheels in assets/wheels. (Note that the basic install script
31
35
uses [DNF](https://docs.fedoraproject.org/en-US/quick-docs/dnf/), but could be easily extended to other Linux OS.)
32
36
@@ -40,7 +44,7 @@ conda activate forge
40
44
41
45
Optional: By default, the packages installation uses conda. If user wants to install system packages on the target machine instead of conda, they can pass the `--use-sudo` to the installation script: `./script/install.sh --use-sudo`.
42
46
43
-
After install, you can run the following command and should see output confirming GRPO training is running (you need a minimum 3 GPU devices).
47
+
After install, you can run the following command and should see output confirming GRPO training is running (you need a minimum 3 GPU devices):
0 commit comments