Skip to content

Commit 80b48d8

Browse files
committed
update
1 parent 25db5d9 commit 80b48d8

File tree

2 files changed

+10
-1
lines changed

2 files changed

+10
-1
lines changed

README.md

Lines changed: 10 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,15 @@ Implementation of <a href="https://arxiv.org/abs/2212.11972">Recurrent Interface
88

99
The big surprise is that the generations can reach this level of fidelity. Will need to verify this on my own machine
1010

11-
Additionally, we will try adding an extra linear attention on the main branch, in addition to the full self attention on the latents. Self conditioning will also be applied to the non-latent images in pixel-space. Let us see how far we can push this approach.
11+
Additionally, we will try adding an extra linear attention on the main branch as well as self conditioning in the pixel-space.
12+
13+
Update:
14+
15+
<img src="./images/sample.png" width="300px"></img>
16+
17+
*130k steps*
18+
19+
It works but the more I think about the paper, the less excited I am. There are a number of issues with the RIN / ISAB architecture. However I think the new sigmoid noise schedule remains interesting as well as the new concept of being able to self-condition on any hidden state of the network.
1220

1321
## Appreciation
1422

@@ -29,6 +37,7 @@ model = RIN(
2937
dim = 256, # model dimensions
3038
image_size = 128, # image size
3139
patch_size = 8, # patch size
40+
depth = 6, # recurrent depth
3241
num_latents = 128, # number of latents. they used 256 in the paper
3342
latent_self_attn_depth = 4, # number of latent self attention blocks per recurrent step, K in the paper
3443
).cuda()

images/sample.png

442 KB
Loading

0 commit comments

Comments
 (0)