You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: the_well/benchmark/models/tfno/README.md
+22-1Lines changed: 22 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,7 +27,6 @@ tags:
27
27
28
28
[The Well](https://github.com/PolymathicAI/the_well) is a 15TB dataset collection of physics simulations. This model is part of the models that have been benchmarked on the Well.
29
29
30
-
31
30
The models have been trained for a fixed time of 12 hours or up to 500 epochs, whichever happens first. The training was performed on a NVIDIA H100 96GB GPU.
32
31
In the time dimension, the context length was set to 4. The batch size was set to maximize the memory usage. We experiment with 5 different learning rates for each model on each dataset.
33
32
We use the model performing best on the validation set to report test set results.
@@ -48,4 +47,26 @@ For benchmarking on the Well, we used the following parameters.
48
47
| Blocks | 4 |
49
48
| Hidden Size| 128 |
50
49
50
+
## Trained Model Versions
51
+
52
+
Below is the list of checkpoints available for the training of TFNO on different datasets of the Well.
Copy file name to clipboardExpand all lines: the_well/benchmark/models/unet_classic/README.md
+22-1Lines changed: 22 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,7 +27,6 @@ tags:
27
27
28
28
[The Well](https://github.com/PolymathicAI/the_well) is a 15TB dataset collection of physics simulations. This model is part of the models that have been benchmarked on the Well.
29
29
30
-
31
30
The models have been trained for a fixed time of 12 hours or up to 500 epochs, whichever happens first. The training was performed on a NVIDIA H100 96GB GPU.
32
31
In the time dimension, the context length was set to 4. The batch size was set to maximize the memory usage. We experiment with 5 different learning rates for each model on each dataset.
33
32
We use the model performing best on the validation set to report test set results.
@@ -50,4 +49,26 @@ For benchmarking on the Well, we used the following parameters.
50
49
| Up/Down Blocks | 4 |
51
50
| Bottleneck Blocks | 1 |
52
51
52
+
## Trained Model Versions
53
+
54
+
Below is the list of checkpoints available for the training of U-Net on different datasets of the Well.
0 commit comments