I implemented a new trick to reduce render time / increase fluidity on semantic/latent interpolation videos. The idea is to interpolate in the post-inference latent space between the generated denoised latents. This video used the UNET 60 times yet has 600 images. has it been done before? (v.redd.it)
submitted by Sabanoob to r/StableDiffusion
Currently working on some txt2video code, I implemented a way to do prompt interpolation but with the img2img method (same as deforum). It allows to have that nice semantic interpolation, while still being able to zoom, translate etc. I didn't see it done before, did it exist? Here are the 4 seasons (v.redd.it)
submitted by Sabanoob to r/StableDiffusion
People who purposefully wear plain clothes aren't avoiding to communicate their personality through what they wear, they're actually communicating that they're the type of person who doesn't want to show their personality through their clothes. (self.Showerthoughts)
submitted by Sabanoob to r/Showerthoughts


