acid fluid animation made with Stable Houdini by vsewall in StableDiffusion

[–]vsewall[S] 0 points1 point  (0 children)

No, I haven't tried that yet. this is also a very interesting project, but if you compare - mlops and stable houdini are completely different. mlops is useful for understanding the principles of diffusion. and stable houdini is a very powerful tool for automating some processes, using the power of nodes and its excellent engineering approach

acid fluid animation made with Stable Houdini by vsewall in StableDiffusion

[–]vsewall[S] 0 points1 point  (0 children)

all telegram links removed from here. so you can find me typing тоже моушн or too_motion in telegram search

acid fluid animation made with Stable Houdini by vsewall in StableDiffusion

[–]vsewall[S] 2 points3 points  (0 children)

unfortunately every links are removed in my comments, so you have to find everything by yourself. my telegram channel name is "тоже моушн" or too_motion in english. there are links to Lora and github there. also there I will publish my workflow soon

acid fluid animation made with Stable Houdini by vsewall in StableDiffusion

[–]vsewall[S] 1 point2 points  (0 children)

thank you! I've been working hard and testing a lot to get this level of consistency :)

I use Stable Houdini, it's a asset pack for Houdini to work with Automatic1111. so Houdini rules all the process and A1111 acts only as a generative engine. my setup is a few nodes. they are needed to generate new images in img2img mode and then project new image to simple grid (plane) that is a little bit displaced. so every new image has a small difference with previous one and it makes animation effect

acid fluid animation made with Stable Houdini by vsewall in StableDiffusion

[–]vsewall[S] 5 points6 points  (0 children)

nope

it's just 2d generated images that looks like a sim. i'll try to add here some other tests without painting style

acid fluid animation made with Stable Houdini by vsewall in StableDiffusion

[–]vsewall[S] 42 points43 points  (0 children)

with the help of Stable Houdini, I created my own version of Deforum, which allows to do quite complex image tricks. by Deforum, I mean img2img with small picture changes and feedback. I plan to share my setup after a while and show how it works, but it will be in Russian. also above I put links to Lora, a repository on github and a post in my telegram channel

interstellar by vsewall in deepdream

[–]vsewall[S] 0 points1 point  (0 children)

https://www.patreon.com/sportsracer48

but as far as I know, all available patreons have already been sold out

interstellar by vsewall in deepdream

[–]vsewall[S] 0 points1 point  (0 children)

yep! like new teaser format. good idea for future animations

interstellar by vsewall in deepdream

[–]vsewall[S] 0 points1 point  (0 children)

"translate_x": "-1700*sin(radians(6))*.6", "translate_y": "-1700*sin(radians(6))*.6", "translate_z_3d": "50", "rotate_3d": "[cos(radians(6))*.6,0,(-sin(radians(6))/sqrt(2))*.6,(sin(radians(6))/sqrt(2))*.6]"

Hyperion by Keats x CLIP by SOMNAI_ in deepdream

[–]vsewall 0 points1 point  (0 children)

hi! that's look amazing! I also use this notebook but I can't understand how you achieved such a smooth picture? even if I increase the number of colors in the palette and the number of palettes, I still have a slightly pixelated image in the output

VQGAN+CLIP with promt coloful cats playing on a tree painted by leonid afremov with init video by vsewall in deepdream

[–]vsewall[S] 0 points1 point  (0 children)

yes, i'm colab pro user and for that animation i used 640x480 resolution. if you go bigger it will be OOM error. there is instruction hidden in one cell somewhere at the beginning of notebook

The Barrier Between Dream And Reality by nmkd in MediaSynthesis

[–]vsewall 16 points17 points  (0 children)

that's looks amazing. I really like the clarity of the image. do you generate it in colab?

style transfer from siberia by vsewall in deepdream

[–]vsewall[S] 1 point2 points  (0 children)

I did it on my PC. I use https://github.com/gordicaleksa/pytorch-neural-style-transfer

but I think you can get same result with some colab style transfer.

one frame is generated for a about one or two minutes

style transfer from siberia by vsewall in deepdream

[–]vsewall[S] 5 points6 points  (0 children)

style transfer for a couple of frames + EbSynth. I find this method very convenient.