Skip to content

Commit 2951e05

Browse files
authored
Merge pull request #84 from jswistak/master
feat: presentation 13th Jan 2025 Controlling the World by Sleight of …
2 parents 72943c3 + 40fe904 commit 2951e05

File tree

5 files changed

+10
-5
lines changed

5 files changed

+10
-5
lines changed

2024/2024_04_22_TIME_Text-To-Image_For_Counterfactual_Explanations/README.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,5 +5,4 @@ This task poses several challenging characteristics for image classification. It
55

66
In next week's presentation, I will talk about the first black-box method of generating Counterfactual Explanations for image classification. TIME (Text-to-Image Models for Counterfactual Explanations) is a novel method that generates CEs using diffusion models combining two different ideas - textual inversion, and EDICT- which create an ingenious method of explaining the image classifiers
77

8-
9-
Presentation will is based on [this paper](https://arxiv.org/pdf/2309.07944.pdf)
8+
Presentation is based on [this paper](https://arxiv.org/pdf/2309.07944.pdf)

2024/2024_05_06_GLOBE-CE/README.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,5 +2,4 @@
22

33
The major drawback of counterfactual explanations lies in their incapacity to extend beyond the local or individual level. Although some research explores the concept of a broader explanation, only a few offer frameworks that are both dependable and computationally manageable. On the seminar on Monday we will be discussing the paper proposing Global & Efficient Counterfactual Explanations (GLOBE-CE), a flexible framework that tackles the reliability and scalability issues associated with current state-of-the-art. The authors also provide unique mathematical analysis of categorical feature translations and experimental evaluation showing that GLOBE-CE performs significantly better than the current methods across multiple metrics.
44

5-
6-
Presentation will is based on [this paper](https://icml.cc/virtual/2023/poster/23706)
5+
Presentation is based on [this paper](https://icml.cc/virtual/2023/poster/23706)
Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,7 @@
1+
# Controlling the World by Sleight of Hand
2+
3+
## Abstract
4+
5+
Humans naturally build mental models of object interactions and dynamics, allowing them to imagine how their surroundings will change if they take a certain action. While generative models today have shown impressive results on generating/editing images unconditionally or conditioned on text, current methods do not provide the ability to perform object manipulation conditioned on actions, an important tool for world modeling and action planning. Therefore, we propose to learn an action-conditional generative models by learning from unlabeled videos of human hands interacting with objects. The vast quantity of such data on the internet allows for efficient scaling which can enable high-performing action-conditional models. Given an image, and the shape/location of a desired hand interaction, CosHand, synthesizes an image of a future after the interaction has occurred. Experiments show that the resulting model can predict the effects of hand-object interactions well, with strong generalization particularly to translation, stretching, and squeezing interactions of unseen objects in unseen environments. Further, CosHand can be sampled many times to predict multiple possible effects, modeling the uncertainty of forces in the interaction/environment. Finally, method generalizes to different embodiments, including non-human hands, i.e. robot hands, suggesting that generative video models can be powerful models for robotics.
6+
7+
Presentation is based on [this paper](https://arxiv.org/pdf/2408.07147)

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ Join us at https://meet.drwhy.ai.
2020
* 02.12 - [Null-text Inversion for Editing Real Images using Guided Diffusion Models](https://github.com/MI2DataLab/MI2DataLab_Seminarium/tree/master/2024/2024_12_02_Null_text_optimization_for_editing_real_images) - Dawid Płudowski
2121
* 09.12 - Diffusion-Based Scene Graph to Image Generation with Masked Contrastive Pre-Training - Tymoteusz Kwieciński
2222
* 16.12 - [Rethinking Visual Counterfactual Explanations Through Region Constraint](https://github.com/MI2DataLab/MI2DataLab_Seminarium/tree/master/2024/2024_12_16_rethinking_visual_counterfactual_explanations_through_region_constraint) - Bartek Sobieski
23-
* 13.01 - Controlling The World by Sleight of Hand - Jakub Świstak
23+
* 13.01 - [Controlling The World by Sleight of Hand](https://github.com/MI2DataLab/MI2DataLab_Seminarium/tree/master/2025/2025_01_13_controlling_the_world_by_sleight_of_hand) - Jakub Świstak
2424
* 20.01 - Connecting counterfactual and attributions modes of explanation - Jan Jakubik
2525

2626
## Year 2023/2024

0 commit comments

Comments
 (0)