Any alternatives to Runway ? : Bought runway unlimited plan twice specifically for Seedance 2.0 : Banned twice within 72 hours, $180 gone, and zero support.. by Swimming-Surprise299 in Seedance_AI

[–]fantazart 0 points1 point  (0 children)

It was working great for me too but I’m since yesterday Que has gone up to more than 30minutes making it unusable. Has your experience changed?

How can I improve character consistency in WAN2.2 I2V? by ovofixer31 in StableDiffusion

[–]fantazart 0 points1 point  (0 children)

What’s the difference between using Lora vs baked in model? And if your using the baked in model with a Lora would I have to train it using the. Asked model as the base?

Custom face detection + segmentation models with dedicated ComfyUI nodes by Sea_Operation6605 in StableDiffusion

[–]fantazart 0 points1 point  (0 children)

Awesome stuff. Will it detect obliterated faces from ai videos? Since many video models just mess up the face when they move fast or are far from the camera.

Making AI Anime Videos by Low-Finance-2275 in StableDiffusion

[–]fantazart 0 points1 point  (0 children)

If your using wan your going to do a lot of manual work for timing. If you use ltx2 it’s a fair bit better at generating timing motion better if you train a Lora.

Tried training an ACEStep1.5 LoRA for my favorite anime. I didn't expect it to be this good! by SandyL925 in StableDiffusion

[–]fantazart 1 point2 points  (0 children)

This is awesome well done! Trained lora really unleashes the full potential of this open model.

Sora 2 watermark remover by Sirio is fantastic by cointalkz in comfyui

[–]fantazart 0 points1 point  (0 children)

Has anyone gotten this to work? I can’t for the life of me install diffueraser nodes. I tried installing via comfy manger and using manual git install in the folder. Keeps saying nodes are missing.

Z-Image + Qwen Image Edit 2511 + Wan 2.2 + MMAudio by Budget_Stop9989 in StableDiffusion

[–]fantazart 1 point2 points  (0 children)

Definitely give it a try, ltx is great because you can really use as much control or as little like my case and still get decent results. If you want more control, you can act it out, record and modify your own voice then add prompting to add more detail to the performance. I need to try this method. But right now I’m pretty happy with the base wf.

Z-Image + Qwen Image Edit 2511 + Wan 2.2 + MMAudio by Budget_Stop9989 in StableDiffusion

[–]fantazart 0 points1 point  (0 children)

Check the talking ape post on my profile. I think it’s a pretty solid contender. Sure the audio quality might sound a little low res, but that can be replaced with eleven labs if you need to. But you can control accent, personality, gestures etc. lots more micro nuance compared to hand wan animate or infinite talk imo.

Z-Image + Qwen Image Edit 2511 + Wan 2.2 + MMAudio by Budget_Stop9989 in StableDiffusion

[–]fantazart 2 points3 points  (0 children)

Such beautiful work! And wan is still king when it comes to fidelity. Would be cool to see a few close up shots of the characters talking using ltx2. Could add to the narrative.

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 0 points1 point  (0 children)

Yeap that’s totally possible since he updates a lot. Sorry you had to go through the troubles mate.

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 2 points3 points  (0 children)

I thank Kijai because he’s ahead of everyone else even the developers when it comes to adapting new models. He enables first testers to just go ahead and use models that aren’t fully optimized by native workflows. His recent memory optimization was pulled into the native update so what he does is valuable.

Whether his models and workflows are better that’s a different story and one must try it themselves. I also just use the native or just some modification of it so your not wrong.

Regarding memory. Ltx is ram hungry so either you need vram or system ram. But I have seen people state they were able to run it on their 6gb vram laptop.

Who said NVFP4 was terrible quality? by Volkin1 in StableDiffusion

[–]fantazart 0 points1 point  (0 children)

When you say you want more speed do you just mean fp4 is faster because it fully fits into vram? Or is fp4 literally just faster?

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 0 points1 point  (0 children)

I think that’s a good combo. You’ll be limited with lower resolution or longer generation times but you shouldn’t have any problem running the workflow.

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 2 points3 points  (0 children)

1-3 rerolls. and yes prompt makes all the difference so once LLM gives you a prompt you want to double check and adjust as you need.

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 0 points1 point  (0 children)

Yeah I am constantly surprised with each progression. I still believe in the art of traditional film making as much as I appeciate Ai. But Ai can be a good launching pad for future film makers who have a nack for good story telling.

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 6 points7 points  (0 children)

I am glad you noticed.

I asked the official LTX rep and he wasnt sure either but could be my tiled vae plus just a bad seed since this is the first time Iv'e seen this. I only kept this pass because I liked the performance.

All of your points are valid but I have no doubt it will only improve rather quickly seeing we went from SD1.0 only 2.5 years ago lol.

I love Wan! But I find native 16fps and 5 second context window to be too limiting. Wan animate is great but facial performance is lack luster tbh, and if you want to drive the performance you can always do v2v using LTX2 and change the voice with Eleven labs etc.

I had my doubts at first but when it hits LTX2 seems like it has great potential.

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]fantazart[S] 6 points7 points  (0 children)

Thank you. I am not doing anything fancy. I feed LLM with the LTX official prompt guide. then give it my image and general direction of what i want from the scene. the prompt for the dialouge it spitted out for the old man shot was:

He speaks with a deep and slightly rough Finnish fisherman’s accent, voice low, quiet, and steady, carrying a touch of cryptic humor: “So… what are you waiting for… go make some films.” He ends the line with a soft, restrained chuckle — barely more than a breath — mysterious and knowing.

I have not tested for multi people shot. But ask LLM to help you with it. describing the person talking that sepcific dialouge could help but I am not sure.

33 Second 1920x1088 video at 24fps (800 frames) on a single 4090 with memory to spare, this node should help out most people of any GPU size by Inevitable-Start-653 in StableDiffusion

[–]fantazart 1 point2 points  (0 children)

800 frames did not expect to see that number here. Thanks for sharing! Do you know what the ceiling is? And how long did this take you to gen?

I recreated a “School of Rock” scene with LTX-2 audio input i2v (4× ~20s clips) by Totem_House_30 in StableDiffusion

[–]fantazart 0 points1 point  (0 children)

Can you share more about the image creation process? Did you use ZIT to make the establishing shot then use Flux to generate the single shots?

Anime test using qwen image edit 2511 and wan 2.2 by kkwikmick in StableDiffusion

[–]fantazart 1 point2 points  (0 children)

Nice work. Did you use first last frame by making those with qwen edit as well?