You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+15-19Lines changed: 15 additions & 19 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@
3
3
4
4
# torchtitan
5
5
6
-
`torchtitan` is currently in a pre-release state and under extensive development.
6
+
`torchtitan` is currently in a pre-release state and under extensive development. Currently we showcase pre-training **Llama 3.1**, **Llama 3**, and **Llama 2** LLMs of various sizes from scratch. To use the latest features of `torchtitan`, we recommend latest PyTorch nightly.
7
7
8
8
`torchtitan` is a proof-of-concept for Large-scale LLM training using native PyTorch. It is (and will continue to be) a repo to showcase PyTorch's latest distributed training features in a clean, minimal codebase. torchtitan is complementary to and not a replacement for any of the great large-scale LLM training codebases such as Megatron, Megablocks, LLM Foundry, Deepspeed, etc. Instead, we hope that the features showcased in torchtitan will be adopted by these codebases quickly. torchtitan is unlikely to ever grow a large community around it.
9
9
@@ -26,34 +26,30 @@ You may want to see how the model is defined or how parallelism techniques are a
26
26
*[torchtitan/parallelisms/pipeline_llama.py](torchtitan/parallelisms/pipeline_llama.py) - helpers for applying Pipeline Parallel to the model
27
27
*[torchtitan/checkpoint.py](torchtitan/checkpoint.py) - utils for saving/loading distributed checkpoints
28
28
*[torchtitan/float8.py](torchtitan/float8.py) - utils for applying Float8 techniques
29
-
*[torchtitan/models/llama/model.py](torchtitan/models/llama/model.py) - the Llama model definition (shared for Llama2 and Llama3 variants)
30
-
31
-
## Pre-Release Updates:
32
-
#### (4/25/2024): `torchtitan` is now public but in a pre-release state and under development.
33
-
Currently we showcase pre-training **Llama 3 and Llama 2** LLMs of various sizes from scratch. `torchtitan` is tested and verified with the PyTorch nightly version `torch-2.4.0.dev20240412`. (We recommend latest PyTorch nightly).
29
+
*[torchtitan/models/llama/model.py](torchtitan/models/llama/model.py) - the Llama model definition (shared for Llama 2 and Llama 3 variants)
8.[Float8](https://discuss.pytorch.org/t/distributed-w-torchtitan-enabling-float8-all-gather-in-fsdp2/209323) support ([how-to](docs/float8.md))
45
41
9.`torch.compile` support
46
-
10. All options easily configured via [toml files](train_configs/)
47
-
11.[Interoperable checkpoints](docs/checkpoint.md) which can be loaded directly
42
+
10. DDP and HSDP
43
+
11. All options easily configured via [toml files](train_configs/)
44
+
12.[Interoperable checkpoints](docs/checkpoint.md) which can be loaded directly into [`torchtune`](https://github.com/pytorch/torchtune) for fine-tuning
48
45
49
-
We report our [Performance](docs/performance.md) verified on 64 A100 GPUs
46
+
We report our [Performance](docs/performance.md) verified on 64/128 GPUs.
50
47
51
48
52
49
### Coming soon
53
50
54
-
1. Context Parallel
55
-
2. Pipeline Parallel (and 3D parallellism)
56
-
3. HSDP
51
+
- Pipeline Parallel (and 3D parallellism)
52
+
- Context Parallel
57
53
58
54
59
55
## Installation
@@ -74,10 +70,10 @@ Once you have confirmed access, you can run the following command to download th
74
70
```bash
75
71
# Get your HF token from https://huggingface.co/settings/tokens
0 commit comments