You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
BayesFlow is a Python library for simulation-based **Amortized Bayesian Inference** with neural networks.
8
-
It provides users with:
7
+
BayesFlow 2 is a Python library for simulation-based **Amortized Bayesian Inference** with neural networks.
8
+
It provides users and researchers with:
9
9
10
10
- A user-friendly API for rapid Bayesian workflows
11
11
- A rich collection of neural network architectures
12
-
- Multi-Backend Support via [Keras3](https://keras.io/keras_3/): You can use [PyTorch](https://github.com/pytorch/pytorch), [TensorFlow](https://github.com/tensorflow/tensorflow), or [JAX](https://github.com/google/jax)
12
+
- Multi-backend support via [Keras3](https://keras.io/keras_3/): You can use [PyTorch](https://github.com/pytorch/pytorch), [TensorFlow](https://github.com/tensorflow/tensorflow), or [JAX](https://github.com/google/jax)
13
13
14
14
BayesFlow is designed to be a flexible and efficient tool that enables rapid statistical inference
15
15
fueled by continuous progress in generative AI and Bayesian inference.
@@ -29,11 +29,35 @@ neural networks for parameter estimation, model comparison, and model validation
29
29
when working with intractable simulators whose behavior as a whole is too
30
30
complex to be described analytically.
31
31
32
-
## Disclaimer
32
+
## Getting Started
33
33
34
-
This is the current dev version of BayesFlow, which constitutes a complete refactor of the library built on Keras 3. This way, you can now use any of the major deep learning libraries as backend for BayesFlow. The refactor is still work in progress with some of the advanced features not yet implemented. We are actively working on them and promise to catch up soon.
34
+
Using the high-level interface is easy, as demonstrated by the minimal working example below:
35
35
36
-
If you encounter any issues, please don't hesitate to open an issue here on [Github](https://github.com/bayesflow-org/bayesflow/issues) or ask questions on our [Discourse Forums](https://discuss.bayesflow.org/).
Check out some of our walk-through notebooks below. We are actively working on porting all notebooks to the new interface so more will be available soon!
7.[Simple model comparison example (One-Sample T-Test)](examples/One_Sample_TTest.ipynb)
103
-
8. More coming soon...
118
+
If you encounter any issues, please don't hesitate to open an issue here on [Github](https://github.com/bayesflow-org/bayesflow/issues) or ask questions on our [Discourse Forums](https://discuss.bayesflow.org/).
104
119
105
120
## Documentation \& Help
106
121
@@ -110,9 +125,8 @@ Documentation is available at https://bayesflow.org. Please use the [BayesFlow F
110
125
111
126
You can cite BayesFlow along the lines of:
112
127
113
-
- We approximated the posterior with neural posterior estimation and learned summary statistics (NPE; Radev et al., 2020), as implemented in the BayesFlow software for amortized Bayesian workflows (Radev et al., 2023a).
114
-
- We approximated the likelihood with neural likelihood estimation (NLE; Papamakarios et al., 2019) without hand-crafted summary statistics, as implemented in the BayesFlow software for amortized Bayesian workflows (Radev et al., 2023b).
115
-
- We performed simultaneous posterior and likelihood estimation with jointly amortized neural approximation (JANA; Radev et al., 2023a), as implemented in the BayesFlow software for amortized Bayesian workflows (Radev et al., 2023b).
128
+
- We approximated the posterior using neural posterior estimation (NPE) with learned summary statistics (Radev et al., 2020), as implemented in the BayesFlow framework for amortized Bayesian inference (Radev et al., 2023a).
129
+
- We approximated the likelihood using neural likelihood estimation (NLE) without hand-crafted summary statistics (Papamakarios et al., 2019), leveraging its implementation in BayesFlow for efficient and flexible inference.
116
130
117
131
1. Radev, S. T., Schmitt, M., Schumacher, L., Elsemüller, L., Pratz, V., Schälte, Y., Köthe, U., & Bürkner, P.-C. (2023a). BayesFlow: Amortized Bayesian workflows with neural networks. *The Journal of Open Source Software, 8(89)*, 5702.([arXiv](https://arxiv.org/abs/2306.16015))([JOSS](https://joss.theoj.org/papers/10.21105/joss.05702))
118
132
2. Radev, S. T., Mertens, U. K., Voss, A., Ardizzone, L., Köthe, U. (2020). BayesFlow: Learning complex stochastic models with invertible neural networks. *IEEE Transactions on Neural Networks and Learning Systems, 33(4)*, 1452-1466. ([arXiv](https://arxiv.org/abs/2003.06281))([IEEE TNNLS](https://ieeexplore.ieee.org/document/9298920))
@@ -153,6 +167,10 @@ You can cite BayesFlow along the lines of:
153
167
}
154
168
```
155
169
170
+
## Awesome Amortized Inference
171
+
172
+
If you are interested in a curated list of resources, including reviews, software, papers, and other resources related to amortized inference, feel free to explore our [community-driven list](https://github.com/bayesflow-org/awesome-amortized-inference).
173
+
156
174
## Acknowledgments
157
175
158
176
This project is currently managed by researchers from Rensselaer Polytechnic Institute, TU Dortmund University, and Heidelberg University. It is partially funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation, Project 528702768). The project is further supported by Germany's Excellence Strategy -- EXC-2075 - 390740016 (Stuttgart Cluster of Excellence SimTech) and EXC-2181 - 390900948 (Heidelberg Cluster of Excellence STRUCTURES), as well as the Informatics for Life initiative funded by the Klaus Tschira Foundation.
0 commit comments