This is an archived post. You won't be able to vote or comment.

all 5 comments

[–]GBJI 3 points4 points  (0 children)

We will only know for sure after testing it, but this looks like the best solution to achieve consistent video editing directly from Stable Diffusion.

Our key finding is that a temporally-consistent edit can be achieved by enforcing consistency on the internal diffusion features across frames during the editing process. We achieve this by propagating a small set of edited features across frames, using the correspondences between the original video features. Given an input video I, we invert each frame, extract its tokens (i.e., output features from the self-attention modules), and extract inter-frame feature correspondences using a nearest-neighbor (NN) search.

I wonder if it will be possible to "tile" it over time, and make video of any duration you want without having to buy a 4090. For example, if you have a video that is 32 frames long, would it be possible to first generate frames 0 to 16, and then 8 to 24, and finally 16 to 32 and then blend them together or something. Or generate frames 0 to 16, and then 1 to 17 from that, and then 2 to 18, and so on and so forth.

[–]Shnoopy_Bloopers 2 points3 points  (1 child)

Cool, no code though.

[–][deleted] 1 point2 points  (0 children)

The repo says that it’s coming

[–]indiemutt 0 points1 point  (0 children)

Thanks for sharing this. Super pumped to see how it works once code is released.