You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For basic usage, this README should give you everything you need to know.
13
-
For deeper insights, you can find the documentation at: https://data-science-in-mechanical-engineering.github.io/mixed_precision_for_JAX/ and our paper at: TODO
13
+
For deeper insights, you can read the [documentation](https://data-science-in-mechanical-engineering.github.io/mixed_precision_for_JAX/) (https://data-science-in-mechanical-engineering.github.io/mixed_precision_for_JAX/) and our [paper](https://www.arxiv.org/pdf/2507.03312) (https://www.arxiv.org/pdf/2507.03312).
14
14
15
15
## Introduction
16
16
@@ -260,6 +260,21 @@ class MultiHeadAttentionBlock(eqx.Module):
260
260
return outputs
261
261
```
262
262
263
+
## Citation
264
+
265
+
To cite this repository, please cite our [paper](https://www.arxiv.org/pdf/2507.03312):
266
+
267
+
```
268
+
@ARTICLE{2025arXiv250703312G,
269
+
author = {{Gr{\"a}fe}, Alexander and {Trimpe}, Sebastian},
270
+
title = "{MPX: Mixed Precision Training for JAX}",
271
+
journal = {arXiv e-prints},
272
+
year = 2025,
273
+
doi = {10.48550/arXiv.2507.03312},
274
+
}
275
+
276
+
277
+
```
263
278
264
279
## Acknowledgements
265
280
We want to thank Partick Kidger for providing equinox and google DeepMind for providing JMP, which was the base for this implementation.
0 commit comments