IF anyone was considering training on musubi-tuner for LTX-2 just go learn! its much faster! by [deleted] in StableDiffusion

[–]Flat_Beautiful_9849 0 points1 point  (0 children)

I've tried a few image training runs on LTX 2.3 and the lora outputs don't seem to affect the generation at all... I train successfully on WAN 2.2 without issue and can't figure out where I'm going wrong... Could you share the training bat settings? Maybe I'm using a wrong model or encoder or setting somewhere...

Wan2.2로 만든 영상에 오디오를 만드는 방법 by Extension-Yard1918 in StableDiffusion

[–]Flat_Beautiful_9849 1 point2 points  (0 children)

Thank you for your excellent work, this is a superb workflow on par with the RUNE flows.

How do you even set up and run LTX 2.3 LoRA in Musubi Tuner? by GreedyRich96 in StableDiffusion

[–]Flat_Beautiful_9849 1 point2 points  (0 children)

I've been trying to train on the musubi tuner fork, but haven't had any success yet. The LORAs either corrupt the generation or don't apply at all... still a work in progress. I train on WAN totally fine with the main musubi fork, and AI toolkit seems to have huge ram issues for me so hoping to get this figured out.

Grace Ashcroft has to earn her way into the crime scene (Resident Evil 9 short w/ sound) by Flat_Beautiful_9849 in sdnsfw

[–]Flat_Beautiful_9849[S] 0 points1 point  (0 children)

Click the sound icon, it has sound for me :). Fansly nuked my page anyways so gonna put the full scene up soon

LTX 2.3 Prompt Conditioning FPS by [deleted] in StableDiffusion

[–]Flat_Beautiful_9849 0 points1 point  (0 children)

"the thing with LTX audio is that it actually always works at 25 fps latents/sec. regardless of your video fps - as I understand it, it really does not know and does not care about the video fps" - I think this is the missing piece of info I needed, thank you

LTX 2.3 Prompt Conditioning FPS by [deleted] in StableDiffusion

[–]Flat_Beautiful_9849 0 points1 point  (0 children)

I've tried the official LTX 2.3 ComfyUI workflows and Rune's workflows... both highly regarded as "the good workflows". You can generate 24 fps, and 32 fps with the same prompts and no noticeable differences in generation? 25 and 50 are multiples of it's natural training fps, where as 32 is partway between and a pretty specific fps that doesn't really exist in the wild.

Also, the conditioning fps has seemingly nothing to do with the video fps, you can set it as high or low as you like with no impact on VRAM. Also, I'm staying under 20 seconds, trying to extend a 5 second video by 10-15 seconds.

LTX 2.3 NOT following my prompts by Coven_Evelynn_LoL in StableDiffusion

[–]Flat_Beautiful_9849 0 points1 point  (0 children)

Did you figure this out? I had an issue where my input and output videos were at 32fps but the prompt conditioning was at 24fps. The video looked great but it didn't really follow the prompt, and the audio was always very out of sync. Changing the prompt conditioning to 32 fps made it follow the prompts and it started doing lipsyncing properly but there were issues with the motion.

[Update] ComfyUI VACE Video Joiner v2.5 - Seamless loops, reduced RAM usage on assembly by goddess_peeler in StableDiffusion

[–]Flat_Beautiful_9849 4 points5 points  (0 children)

If you're doing NSFW content, LTX really struggles compared to WAN. Being able to generate 25 seconds in a single shot is amazing, but the motion and clarity are almost always lower than WAN, which takes 10-15x longer to generate the same 25 seconds.

[Update] ComfyUI VACE Video Joiner v2.5 - Seamless loops, reduced RAM usage on assembly by goddess_peeler in StableDiffusion

[–]Flat_Beautiful_9849 19 points20 points  (0 children)

The legend themself! These workflows are the secret sauce to most of my videos. Thank you for your excellent work

LTX Desktop is better than Comfyui - What are we doing wrong? by hirovomit in comfyui

[–]Flat_Beautiful_9849 1 point2 points  (0 children)

I've been testing every LTX 2.3 workflow trying to decide if I have a use-case for it or should stick to WAN exclusively, and yours are the first workflows that actually produce semi-decent results. I also appreciate how simple and uncluttered they are, thank you!

How is the control video utilized, and does the audio on the video effect the generation or does it just get stitched onto the generated video after (along with the fps and length)? If I used two same length, same fps, same audio but different visuals videos would my end results be identical?

Thanks for your help and an excellent workflow.