You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
<imgsrc="static/images/mm_envs.png" alt="Simulation and hardware environments">
321
320
</figure>
322
321
<p>
323
-
Experiment Testbeds. In simulation (left) and hardware (right) we control data from two sensors: RGB and infrared (IR) camera. In our controlled experiments, the ground-truth safety-relevant state variable is heat, which is more observable from the IR data than the RGB.
322
+
We evaluate latent safety filters in both simulation and hardware environments designed to reveal how partial observability impacts safe behavior.
323
+
</p>
324
+
<p>
325
+
In simulation, we introduce the <strong>thermal unicycle</strong>, a Dubins-style unicycle model augmented with a latent heat variable that increases as the agent approaches a heat source. The agent receives either RGB or infrared (IR) images and must prevent overheating. This setup is intentionally simple and controllable, allowing us to isolate how safety filters behave when safety-relevant features are only partially observable.
326
+
</p>
327
+
<p>
328
+
On hardware, we use a <strong>Franka Research 3 manipulator</strong> heating a pot of wax. During training, the robot observes both RGB and IR data, where the IR modality provides a privileged view of heat, the true safety variable. At test time, only RGB observations are used, enabling us to evaluate how well latent representations trained under different modalities encode or omit safety-critical information in the real world.
<h2class="title is-3">Mutual Information as a Measure of Observability</h2>
338
344
<divclass="content has-text-justified">
345
+
<p>
346
+
Our mutual information (MI) metric quantifies how much uncertainty about safety outcomes is reduced by observing a particular input modality (e.g., RGB or infrared). We compute a Barber-Agakov lower bound on MI between observations and binary safety labels to measure how well each modality captures safety-relevant features. Higher MI indicates that the modality more reliably encodes features necessary for safety prediction.
347
+
</p>
348
+
339
349
<figureclass="image is-centered">
340
-
<imgsrc="static/images/MI_metric.png" alt="MI reveals greater separation than accuracy-based metrics for quantifying the observability of safety constraints from high-D obs" width="1100">
Our mutual information (MI) metric quantifies how much uncertainty about safety outcomes is reduced by observing a particular input modality (e.g., RGB or infrared). We compute a Barber-Agakov lower bound on MI between observations and binary safety labels to measure how well each modality captures safety-relevant features. Higher MI indicates that the modality more reliably encodes features necessary for safety prediction. We find that IR observations exhibit much higher normalized MI than RGB alone, meaning RGB-only models often lack sufficient safety information, which explains their myopic, “avoid seeing failure” behaviors.
353
+
We find that IR observations exhibit much higher normalized MI than RGB alone, meaning RGB-only models often lack sufficient safety information, which explains their myopic, “avoid seeing failure” behaviors. Furthermore, we find that the MI metrics are more indicative than traditional metrics, such as accuracy and balanced accuracy, when identifying degenerative latent states: our simulation balanced accuracy could potentially indicate a sufficient classifier, even though the RGB channel is designed to provide little to no indication regarding failure.
345
354
</p>
346
-
347
355
</div>
348
356
</div>
349
357
</div>
@@ -359,13 +367,32 @@ <h2 class="title is-3">Mutual Information as a Measure of Observability</h2>
We evaluate the quality of the learned latent representations by examining how well they encode safety-relevant state information. Models trained only on RGB observations often produce latent states that fail to represent temperature, leading to visually correct but unsafe predictions. In contrast, our multimodal approach learns latent states that embed the underlying thermal dynamics, enabling proactive interventions that maintain safety.
372
+
</p>
362
373
<figureclass="image is-centered">
363
374
<imgsrc="static/images/hw_qual_v3.png" alt="RGB-only training is unable to understand safety outcomes of actions" width="1100">
364
375
</figure>
365
376
<p>
366
-
We evaluate the quality of the learned latent representations by examining how well they encode safety-relevant state information. Models trained only on RGB observations often produce latent states that fail to represent temperature, leading to visually correct but unsafe predictions. In contrast, our multimodal-supervised approach—trained with both RGB and IR data but deployed with RGB alone—learns latent states that embed the underlying thermal dynamics, enabling proactive interventions that maintain safety.
377
+
To quantify latent representation quality, we introduce two diagnostic tests. The <em>latent state test</em> measures how much safety-relevant information (e.g., heat) is directly encoded in the learned latent state, while the <em>latent dynamics test</em> evaluates whether the world model's imagined rollouts understand how that information evolves time. Together, these tests reveal whether the learned latent space both contains and maintains the safety features needed for effective control. The multimodal simulation and hardware methods align with our qualitative observations: the latent features degrade when safety-critical features are not directly observable.
367
378
</p>
368
379
</div>
380
+
381
+
<divclass="columns has-text-justified">
382
+
<divclass="column">
383
+
<h3class="title is-5">Latent State Test</h3>
384
+
<figureclass="image is-centered">
385
+
<imgsrc="static/images/latent_state_test.png" alt="RGB-only training is unable to understand safety outcomes of actions" width="1100">
386
+
</figure>
387
+
</div>
388
+
389
+
<divclass="column">
390
+
<h3class="title is-5">Latent Dynamics Test</h3>
391
+
<figureclass="image is-centered">
392
+
<imgsrc="static/images/latent_dynamics_test.png" alt="RGB-only training is unable to understand safety outcomes of actions" width="1100">
The multimodal-supervised safety filter also anticipates overheating and lifts the pan before the wax fails. Trained with RGB + IR data but deployed using only RGB, the controller maintains safety even under partial observability.
485
+
The multimodal-supervised safety filter also anticipates overheating and lifts the pan before failure. Trained with RGB + IR data but deployed using only RGB, the controller maintains safety even under partial observability.
0 commit comments