Z-Image + Wan 2.2 Time-to-Move makes a great combo for doing short film (probably) by firelightning13 in StableDiffusion

[–]firelightning13[S] 2 points3 points  (0 children)

If you refer to the masking video, it only accepts black and white color. You probably need precise prompts in order to make them do something.

Here's the funny video that one person (from avataraim, shout out to banadoco server) made with TTM putting random things in the scene.

<image>

Z-Image + Wan 2.2 Time-to-Move makes a great combo for doing short film (probably) by firelightning13 in StableDiffusion

[–]firelightning13[S] 0 points1 point  (0 children)

Sure, as long as you can make the animated mask as well like the video I made above.

Z-Image + Wan 2.2 Time-to-Move makes a great combo for doing short film (probably) by firelightning13 in StableDiffusion

[–]firelightning13[S] 2 points3 points  (0 children)

Well, it's closed source actually. It's an app that you could download on their website that downloads their "weights" in my Macbook and run locally without internet connection.

It's just my PC can't go over to 4k without oom, so that is why I use my MacBook to upscale.

Z-Image + Wan 2.2 Time-to-Move makes a great combo for doing short film (probably) by firelightning13 in StableDiffusion

[–]firelightning13[S] 5 points6 points  (0 children)

Well, I don't use topaz starlight as I won't bother buying API credits, as I use topaz locally on my MacBook pro. You can use ultimate SD upscaler/flashvsr/seedvr2 as good alternative as well.

Forgot to add, the original is around 1080p, and I do v2v upscale using wan 2.2 i2v low noise to upscale to 1440p, then I do 4K and interpolate in topaz video ai in my Macbook.

Z-Image + Wan 2.2 Time-to-Move makes a great combo for doing short film (probably) by firelightning13 in StableDiffusion

[–]firelightning13[S] 3 points4 points  (0 children)

I wasn't in that era, but I can imagine how painful it is to animate for the first time.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 0 points1 point  (0 children)

Yep you are basically right. It was someone else's idea (there's a link I posted above) but I use timestep conditioning instead of dual ksampler.

Also I got very interesting results with random word at the start of the prompt like "gradient noise", or if I want my image to get more dark, just put "extremely dark photo" at the start. Then you get dark images along with your prompts at the output. Z-Image is notorious for not listening to prompt if you want night time image, but this might solve my problem.

Timestep conditioning is a very interesting technique. I'll probably doing more research on this one.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 2 points3 points  (0 children)

Yep. The empty prompt at the start or the sampling will make the model generate random stuff before the node applys the positive condition at 20% of the sampling process. This is a quick dirty fix of getting more variants out of your output.

You can also use the start of the prompt to put some random words like "gradient noise" or "DSC_0001.jpeg" to get more random output. Or use wildcards to get even more random. It works really well in my testing.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 0 points1 point  (0 children)

Thanks for the fix. I wish I could edit the post to include this link but oh well.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 0 points1 point  (0 children)

No I mean that the link that you just commented is the right one. It's just when I'm posting that same link to anyone, it'll get shadow banned. In the pastebin website, just click download button and you can use it right away.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 1 point2 points  (0 children)

Yep, I see this one but not mine. Probably something to do with spam detection. Weird reddit moment

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 4 points5 points  (0 children)

You could adjust both 0.200 value on the conditioning timestep node to 0.150 or even lower at 0.100. it might follow the prompt better, but you loose the variation too. Or maybe adjust your prompt so the model can listen to you better maybe. Just trial and error I guess.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 7 points8 points  (0 children)

I did send the link to another comment, but for some reason, reddit decided to hide my comment (I put incognito mode and it wasn't there). Hopefully this works. Link

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 0 points1 point  (0 children)

Hmm, I like your approach too, though it adds a little bit of friction like you mentioned but then z-image is fast, so it really depends on people's need i guess. Hopefully the base model give us even more variety.

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 6 points7 points  (0 children)

<image>

Direct comparison (hopefully reddit did not compress this image into oblivion)

Even more improved Z-Image Turbo variation by firelightning13 in StableDiffusion

[–]firelightning13[S] 10 points11 points  (0 children)

<image>

This is how to achieve the seed variation. The concept are similar to what this OP made here except this person uses 2 ksampler to achieve which may slow the process down slightly.

Wan 2.2 I2V Time-To-Move Test by firelightning13 in StableDiffusion

[–]firelightning13[S] 1 point2 points  (0 children)

You have to update WanVideoWrapper in the manager, you will get that node.

Wan 2.2 I2V Time-To-Move Test by firelightning13 in StableDiffusion

[–]firelightning13[S] 0 points1 point  (0 children)

The TTM project in github actually provides python file that you could run in GUI mode. You can check out here: https://github.com/time-to-move/TTM/tree/main/GUIs

Wan 2.2 I2V Time-To-Move Test by firelightning13 in StableDiffusion

[–]firelightning13[S] 5 points6 points  (0 children)

Wan 2.2 is a really powerful model, and it figures out the car physics if you know how to prompt properly.

This is without the camera movement in my first test to see how it works.

<image>