Skip to content

Commit 9dbf2dd

Browse files
committed
Upload blog post 2025-7-21 and update team member list
1 parent 7b24f5e commit 9dbf2dd

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

43 files changed

+925
-36
lines changed

_config.yml

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -380,6 +380,11 @@ team:
380380
website: https://flanusse.net/
381381
bio: Francois Lanusse is an interdisciplinary researcher at the intersection of Deep Learning, Statistical Modeling, and Observational Cosmology. Dr. Lanusse holds a permanent position at the CNRS, and is currently an Associate Research Scientist at the Simons Foundation. He received his PhD in Astrophysics at CEA Paris-Saclay and was subsequently a postdoctoral researcher at Carnegie Mellon University and UC Berkeley.
382382

383+
- full_name: Tanya Marwah
384+
avatar: tanya_marwah.png
385+
website: https://tm157.github.io/
386+
bio: Tanya Marwah is a Research Fellow at the Simons Foundation working with Polymathic AI. She is broadly interested in theoretical and empirical foundations of Machine Learning and its applications to scientific domains. Her current interests are around generative modeling of scientific phenomena, inverse problems and building scientific agents. Her ultimate goal is to develop ML algorithms and methods that help us accelerate the scientific process and enable scientific discovery. She recently graduated with a PhD from the Machine Learning Department at Carnegie Mellon University and holds a Masters in Robotics from the Robotics Institute at CMU and was a Siebel Scholar.
387+
383388
- full_name: Michael McCabe
384389
avatar: michael_mccabe.jpg
385390
website: https://mikemccabe210.github.io/

_posts/2025-07-21-latent-space.md

Lines changed: 80 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,80 @@
1+
---
2+
layout: post
3+
title: "Lost in Latent Space: the Pros and Cons of Latent Physics Emulation"
4+
authors: François Rozet, Ruben Ohana, Michael McCabe, Gilles Louppe, François Lanusse, Shirley Ho
5+
shorttitle: "Lost in Latent Space"
6+
date: 2025-07-21 9:00
7+
smallimage: latent_space_s.jpg
8+
image: latent_space.jpg
9+
blurb: We show that latent diffusion models are robust to compression in the context of physics emulation, reducing computational cost while consistently outperforming non-generative alternatives.
10+
shortblurb: We show that latent diffusion models are robust to compression in the context of physics emulation, reducing computational cost while consistently outperforming non-generative alternatives.
11+
splashimage: /images/blog/latent_space.jpg
12+
link: https://arxiv.org/abs/2507.02608
13+
github_link: https://github.com/PolymathicAI/lola
14+
permalink: /blog/lostinlatentspace/
15+
---
16+
17+
Numerical simulations are fundamental to scientific progress, enabling everything from weather forecasting to plasma control in fusion reactors. However, achieving high-fidelity results often requires significant computational resources, making these simulations a bottleneck for rapid research and development.
18+
19+
At <a href="https://polymathic-ai.org/">Polymathic</a>, we believe that neural network-based emulators are a promising alternative to traditional numerical solvers, enabling orders of magnitude faster simulations. Recently, latent diffusion models were applied with success to the problem of emulating dynamical systems (<a href="https://arxiv.org/abs/2307.10422">Gao et al., 2023</a>; <a href="https://arxiv.org/abs/2403.05940">Du et al., 2024</a>; <a href="https://arxiv.org/abs/2504.18720">Andry et al., 2025</a>), sometimes even outperforming pixel-space emulation. In this work, we asked ourselves a simple question: *What is the impact of latent-space compression on emulation accuracy?*
20+
21+
The answer surprised us, and we think it will surprise you too.
22+
23+
#### From Pixel Space to Latent Space
24+
25+
The core idea of latent diffusion models (<a href="https://arxiv.org/abs/2112.10752">Rombach et al., 2022</a>), which have proven highly effective for image and video generation, is to perform the generative process not in the high-dimensional pixel space, but in a compressed, low-dimensional latent space learned by an autoencoder. For natural images, compression serves a dual purpose: reducing computational cost and filtering out perceptually irrelevant patterns that might distract the generative model from semantically meaningful information.
26+
27+
In our case, the methodology involves three stages. First, an autoencoder is trained to compress high-dimensional physical states into compact latent representations. Second, a diffusion model is trained to predict/emulate the temporal evolution of the system within this compressed latent space. Third, after training, the diffusion model is used to predict the sequence of latent states which are then mapped back to the pixel space with the autoencoder's decoder.
28+
29+
<p align="center">
30+
<img src="/images/blog/latent_emulation.svg" alt="Latent emulation" width="95%" style="mix-blend-mode: darken;">
31+
</p>
32+
33+
#### Findings
34+
35+
To answer our research question, we trained and evaluated latent-space emulators across a wide range of compression rates – from modest (x48) to extreme (x1280) – on three challenging datasets from <a href="https://polymathic-ai.org/blog/thewell">The Well</a>:
36+
37+
- **Euler Multi-Quadrants**, describing compressible fluids and shock waves.
38+
<p align="center">
39+
<video width="95%" controls>
40+
<source src="/images/blog/latent_space_vid/euler_f32c64.mp4" type="video/mp4">
41+
Your browser does not support the video tag.
42+
</video>
43+
</p>
44+
45+
- **Rayleigh-Bénard**, modeling buoyancy driven convection currents.
46+
<p align="center">
47+
<video width="95%" controls>
48+
<source src="/images/blog/latent_space_vid/rb_f32c64.mp4" type="video/mp4">
49+
Your browser does not support the video tag.
50+
</video>
51+
</p>
52+
53+
- **Turbulence Gravity Cooling**, simulating the formation and radiative cooling of stars in interstellar media.
54+
<p align="center">
55+
<video width="95%" controls>
56+
<source src="/images/blog/latent_space_vid/tgc_f32c64.mp4" type="video/mp4">
57+
Your browser does not support the video tag.
58+
</video>
59+
</p>
60+
61+
Our experiments reveal two key findings.
62+
63+
**1. Robustness to Compression**
64+
65+
Our most striking finding is the **remarkable resilience of latent emulation to the compression rate** of the latent space with respect to pixel space. While reconstruction quality deteriorates as compression increases, we do not observe any significant degradation in the emulation accuracy itself. In all cases, **latent emulators outperform pixel-space baselines**, despite using fewer parameters and less training compute.
66+
67+
Nevertheless, our evaluation reveals potential overfitting issues at extreme compression rates. This makes intuitive sense: as compression increases, the effective size of the dataset in latent space decreases, making overfitting more likely at fixed model capacity. This underscores the importance of efforts like <a href="https://polymathic-ai.org/blog/thewell">The Well</a>, which provides curated, large-scale physics data for training and benchmarking emulators.
68+
69+
**2. Generative Models over Deterministic Solvers**
70+
71+
Across all tasks and compression rates, **diffusion-based emulators are consistently more accurate than deterministic neural solvers**. They not only produce better and more plausible trajectories, but also capture the uncertainty and diversity inherent to turbulent and chaotic dynamical systems.
72+
73+
#### Practical Recommendations for Practitioners
74+
75+
Our findings translate into clear, actionable recommendations for practitioners developing physics emulators. First, **try latent-space approaches**. They offer reduced computational requirements and provide comparable or superior performances across a wide range of compression rates. In our case, it also greatly simplified the development and training of the emulator as we could rely on widespread transformer architectures with well known scaling properties. Second, **prefer generative over deterministic emulators**. They yield better accuracy, more plausible dynamics, stable rollouts, and naturally handle uncertainty.
76+
77+
For more details, check out the <a href="https://arxiv.org/abs/2507.02608">paper</a>.
78+
79+
---
80+
Image by [JJ Ying](https://unsplash.com/@jjying?utm_content=creditCopyText&utm_medium=referral&utm_source=unsplash) via [Unsplash](https://unsplash.com/photos/white-cloth-lot-WmnsGyaFnCQ?utm_content=creditCopyText&utm_medium=referral&utm_source=unsplash).

collaborators/david-fouhey.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "David Fouhey"
5+
---
6+
7+
Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "Francesco Pio Ramunno"
5+
---
6+
7+

collaborators/francois-rozet.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "Francois Rozet"
5+
---
6+
7+
Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "Ghazal Khalighinejad"
5+
---
6+
7+

collaborators/helen-qu.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "Helen Qu"
5+
---
6+
7+

collaborators/jake-kovalic.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "Jake Kovalic"
5+
---
6+
7+

collaborators/jiequn-han.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
2+
---
3+
layout: collaborator
4+
full_name: "Jiequn Han"
5+
---
6+
7+

collaborators/keiya-hirashima.md

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
1-
---
2-
layout: collaborator
3-
full_name: "Keiya Hirashima"
4-
---
5-
6-
1+
2+
---
3+
layout: collaborator
4+
full_name: "Keiya Hirashima"
5+
---
6+
7+

0 commit comments

Comments
 (0)