Skip to content

Commit 748140e

Browse files
authored
Merge pull request #66 from AI-Hypercomputer/bvandermoon-tpu-recipes
Bump MaxText recipes to new tpu-recipes version
2 parents 2a33362 + 4541811 commit 748140e

File tree

7 files changed

+28
-28
lines changed

7 files changed

+28
-28
lines changed

training/trillium/GPT3-175B-MaxText/bf16/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -43,7 +43,7 @@ completed step: 15, seconds: 17.182, TFLOP/s/device: 384.891, Tokens/s/device: 3
4343

4444
### Workload Details
4545

46-
For reference, here are the `gpt_3_175b_bf16` workload details as found in `[email protected].0`:
46+
For reference, here are the `gpt_3_175b_bf16` workload details as found in `[email protected].2`:
4747

4848
```
4949
MaxTextModel(
@@ -72,4 +72,4 @@ MaxTextModel(
7272
)
7373
```
7474

75-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.0/benchmarks/maxtext_trillium_model_configs.py#L287) file within the MaxText repository.
75+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.2/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

training/trillium/Llama2-70B-MaxText/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2(https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -43,7 +43,7 @@ completed step: 16, seconds: 9.052, TFLOP/s/device: 402.274, Tokens/s/device: 90
4343

4444
### Workload Details
4545

46-
For reference, here are the `llama2_70b_4096_sc` workload details as found in `[email protected].0`:
46+
For reference, here are the `llama2_70b_4096_sc` workload details as found in `[email protected].2`:
4747

4848
```
4949
MaxTextModel(
@@ -76,4 +76,4 @@ MaxTextModel(
7676
)
7777
```
7878

79-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.0/benchmarks/maxtext_trillium_model_configs.py#L410) file within the MaxText repository.
79+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.2/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

training/trillium/Llama3-8B-MaxText/v6e-8/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -44,7 +44,7 @@ If you would like to run on multiple slices of v6e-8, you may modify the `--num_
4444

4545
### Workload Details
4646

47-
For reference, here are the `llama3_1_8b_8192_no_collective_matmul` workload details as found in `[email protected].0`:
47+
For reference, here are the `llama3_1_8b_8192_no_collective_matmul` workload details as found in `[email protected].2`:
4848

4949
```
5050
MaxTextModel(
@@ -90,4 +90,4 @@ For reference, here are the `llama3_1_8b_8192_no_collective_matmul` workload det
9090
)
9191
```
9292

93-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.0/benchmarks/maxtext_trillium_model_configs.py#L858-L901) file within the MaxText repository.
93+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.2/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

training/trillium/Llama3.1-405B-MaxText/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -43,7 +43,7 @@ completed step: 14, seconds: 54.803, TFLOP/s/device: 392.454, Tokens/s/device: 1
4343

4444
### Workload Details
4545

46-
For reference, here are the `llama3_1_405b_8192_pure_fsdp_ici` workload details as found in `[email protected].0`:
46+
For reference, here are the `llama3_1_405b_8192_pure_fsdp_ici` workload details as found in `[email protected].2`:
4747

4848
```
4949
MaxTextModel(
@@ -76,4 +76,4 @@ MaxTextModel(
7676
)
7777
```
7878

79-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.0/benchmarks/maxtext_trillium_model_configs.py#L767) file within the MaxText repository.
79+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.2/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

training/trillium/Llama3.1-70B-MaxText/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -44,7 +44,7 @@ If you would like to run on multiple slices of v6e-256, you may modify the `--nu
4444

4545
### Workload Details
4646

47-
For reference, here are the `llama3_1_70b_8192` workload details as found in `[email protected].0`:
47+
For reference, here are the `llama3_1_70b_8192` workload details as found in `[email protected].2`:
4848

4949
```
5050
MaxTextModel(
@@ -87,4 +87,4 @@ For reference, here are the `llama3_1_70b_8192` workload details as found in `Ma
8787
)
8888
```
8989

90-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/243b25e480f7550a0c389fa95cd3adcc716fe0df/benchmarks/maxtext_trillium_model_configs.py#L932-L972) file within the MaxText repository.
90+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/243b25e480f7550a0c389fa95cd3adcc716fe0df/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

training/trillium/Mistral-7B-MaxText/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -44,7 +44,7 @@ If you would like to run on multiple slices of v6e-8, you may modify the `--num_
4444

4545
### Workload Details
4646

47-
For reference, here are the `mistral_7b` workload details as found in `[email protected].0`:
47+
For reference, here are the `mistral_7b` workload details as found in `[email protected].2`:
4848

4949
```
5050
MaxTextModel(
@@ -90,4 +90,4 @@ For reference, here are the `mistral_7b` workload details as found in `MaxText@t
9090
)
9191
```
9292

93-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.0/benchmarks/maxtext_trillium_model_configs.py#L1217-L1260) file within the MaxText repository.
93+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.2/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

training/trillium/Mixtral-8x7B-MaxText/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -8,9 +8,9 @@ Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/m
88
### Install MaxText and Build Docker Image
99
Please follow this [link](https://github.com/AI-Hypercomputer/tpu-recipes/blob/main/training/trillium/MAXTEXT_README.md) to install maxtext and build the docker image. The following variables should be set:
1010

11-
In step 1, use the MaxText [tpu-recipes-v0.1.1](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.1) tag to run this recipe:
11+
In step 1, use the MaxText [tpu-recipes-v0.1.2](https://github.com/AI-Hypercomputer/maxtext/releases/tag/tpu-recipes-v0.1.2) tag to run this recipe:
1212
```
13-
git checkout tpu-recipes-v0.1.1
13+
git checkout tpu-recipes-v0.1.2
1414
```
1515

1616
In step 2, use the jax-stable-stack image containing JAX 0.5.2:
@@ -44,7 +44,7 @@ completed step: 11, seconds: 13.484, TFLOP/s/device: 302.311, Tokens/s/device: 3
4444

4545
### Workload Details
4646

47-
For reference, here are the `mixtral_8x7b_dropped` workload details as found in `[email protected].0`:
47+
For reference, here are the `mixtral_8x7b_dropped` workload details as found in `[email protected].2`:
4848

4949
```
5050
MaxTextModel(
@@ -84,4 +84,4 @@ MaxTextModel(
8484
)
8585
```
8686

87-
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.0/benchmarks/maxtext_trillium_model_configs.py#L1296) file within the MaxText repository.
87+
This equivalent workload code can be found in the [maxtext_trillium_model_configs.py](https://github.com/AI-Hypercomputer/maxtext/blob/tpu-recipes-v0.1.2/benchmarks/maxtext_trillium_model_configs.py) file within the MaxText repository.

0 commit comments

Comments
 (0)