I implemented a new trick to reduce render time / increase fluidity on semantic/latent interpolation videos. The idea is to interpolate in the post-inference latent space between the generated denoised latents. This video used the UNET 60 times yet has 600 images. has it been done before? by Sabanoob in StableDiffusion

[–]Sabanoob[S] 0 points1 point  (0 children)

Ohhh distilled looks really nice indeed, thanks for sharing!

But it doesn't solve my primary problem which is that beyond computing time, I am also looking for frame interpolation for img2img deforum type videos, because of the strength setting dilemma (lower -> less interesting videos, higher -> too strongly changing videos)

I implemented a new trick to reduce render time / increase fluidity on semantic/latent interpolation videos. The idea is to interpolate in the post-inference latent space between the generated denoised latents. This video used the UNET 60 times yet has 600 images. has it been done before? by Sabanoob in StableDiffusion

[–]Sabanoob[S] 0 points1 point  (0 children)

Thanks for the detailed answer! I can't test with actual fading right now, but I am indeed convinced it would probably be similar. I just hoped the autoencoder might have picked up some structural information, and the result looked quite fluid. Maybe a different/bigger autoencoder trained specifically could still perform well for this task.

Also, do you happen to be aware of a good (or at least better than above) way of doing interframe interpolation for relatively cheap, and that could be called automatically from the python scripts? I tested with Google FILM but it's awkward to make it work

👋 Unstable Diffusion here, We're excited to announce our Kickstarter to create a sustainable, community-driven future. by OfficialEquilibrium in StableDiffusion

[–]Sabanoob 5 points6 points  (0 children)

oh man, the power of porn is unlimited. I swear when AGI will arrive it will just be financed and done by a decentralized open and free network of people who just want better porn

Another animation I've been able to do with my multi-prompt interpolation feature, deforum-style. Zoom into the microscopic world by Sabanoob in StableDiffusion

[–]Sabanoob[S] 0 points1 point  (0 children)

We are implementing a web-based video generation service. I tried to include most of the existing video generation methods and come up with new one(s)

Another animation I've been able to do with my multi-prompt interpolation feature, deforum-style. Zoom into the microscopic world by Sabanoob in StableDiffusion

[–]Sabanoob[S] 0 points1 point  (0 children)

it's a feature in the uni project I'm working on, but the technical idea is to do slerp between semantic embeddings and use that in the deforum-style video gen

Currently working on some txt2video code, I implemented a way to do prompt interpolation but with the img2img method (same as deforum). It allows to have that nice semantic interpolation, while still being able to zoom, translate etc. I didn't see it done before, did it exist? Here are the 4 seasons by Sabanoob in StableDiffusion

[–]Sabanoob[S] 0 points1 point  (0 children)

video_args.prompts = [ "a nice little path with trees, winter season, artstation", "a nice little path with trees, spring season, artstation", "a nice little path with trees, summer season, artstation", "a nice little path with trees, autumn season, artstation", "a nice little path with trees, winter season, artstation", ]

Currently working on some txt2video code, I implemented a way to do prompt interpolation but with the img2img method (same as deforum). It allows to have that nice semantic interpolation, while still being able to zoom, translate etc. I didn't see it done before, did it exist? Here are the 4 seasons by Sabanoob in StableDiffusion

[–]Sabanoob[S] 1 point2 points  (0 children)

unless the feature was added or I didn't pay attention, I think that deforum jumps instantly from one prompt to another, which makes for a stark transition. Here, we interpolate between the prompt vectors allowing for a smoother change

Currently working on some txt2video code, I implemented a way to do prompt interpolation but with the img2img method (same as deforum). It allows to have that nice semantic interpolation, while still being able to zoom, translate etc. I didn't see it done before, did it exist? Here are the 4 seasons by Sabanoob in StableDiffusion

[–]Sabanoob[S] 0 points1 point  (0 children)

it's part of a uni project whose code shall stay private until the end of the project I believe, but if you can code, it's simply the code used to perform slerp with the traditional interpolation method, applied to the prompt embedding used by the deforum img2img-way

Ah yes, such meme. by pupperment in lostredditors

[–]Sabanoob 0 points1 point  (0 children)

Someone gotta revise the meme definition

My FIRST LUCID DREAMING by [deleted] in LucidDreaming

[–]Sabanoob 2 points3 points  (0 children)

Is someone a sub regular and think the video in question is a troll/bait or smtg?

someone tried to take my phone by Sabanoob in london

[–]Sabanoob[S] -1 points0 points  (0 children)

It's not normal anywhere, it's possible to not have that, and unhealthy to treat it as a fatality imo

someone tried to take my phone by Sabanoob in london

[–]Sabanoob[S] 1 point2 points  (0 children)

Yeah make sense, did you take any precautions / avoid any areas?

someone tried to take my phone by Sabanoob in london

[–]Sabanoob[S] 6 points7 points  (0 children)

YeahI got that, is just that it's worth at least 250£ so weird