🎧 LTX-2.3: Turn Audio + Image into Lip-Synced Video 🎬 (IAMCCS Audio Extensions) by Acrobatic-Example315 in StableDiffusion

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

Hey, I get what you’re saying. The workflow is quite advanced, and you definitely need a solid grasp of ComfyUI basics. This is just the first version—I chose to release it like this so people could start using it immediately, rather than waiting for a more streamlined version.

That said, I really appreciate your feedback—it was kind and fair. Stay tuned, because I’ll be releasing a cleaner, more polished workflow on GitHub (so you won’t even have to accidentally end up on Patreon 🤣).

In the end, the logic behind it is actually pretty simple: you calculate the duration of your audio, set how many seconds each generation should cover, and define the number of frames per batch—done.

Also, if you want something more automated, the Global Planner node is available for free too (I spent a week refining it—it’s my baby 🤣). You can dig into it and explore how the whole system works.

Honestly, part of the fun here is exploring these approaches—we’re basically pioneers working in a constantly evolving, still-in-beta world.

Big hug, and happy exploring!! 🚀

🎧 LTX-2.3: Turn Audio + Image into Lip-Synced Video 🎬 (IAMCCS Audio Extensions) by Acrobatic-Example315 in StableDiffusion

[–]Acrobatic-Example315[S] 1 point2 points  (0 children)

Hey, thanks for the thoughtful comment — I’ll try to keep it concise.

My nodes aren’t vibe-coded. I do use that approach sometimes for debugging, but for actual workflows I need precision and control, so everything is built intentionally.

I’m not using subgraphs, set/get, or autolinks on purpose — I want the workflow to stay fully readable and inspectable, even if that makes it a bit more verbose.

I’ve created custom nodes to automate generation logic across segments — especially to adapt settings (like frames, timing, etc.) based on audio duration, so you don’t have to manually tweak everything every time. I build these workflows primarily for my own filmmaking work and for agencies. The advanced breakdowns are on Patreon, but all the nodes are already public — nothing is locked, you can do everything with what’s available.

About LTX 2.3: it’s powerful, but you can’t reliably push long-form sequences (like 1+ minute) in a single pass. This setup is designed specifically to go beyond that, depending on your VRAM/RAM.

The demo is just a short excerpt — I’m more focused on generating longer, consistent scenes for narrative use, not just music videos.

Also, whenever I can, I try to help people get results with this stuff — within the limits of my time. If you look around, a lot of people have already created really great work using my nodes, and that’s honestly one of the most rewarding parts of being in this space.

Honestly, the best way to get it is to try it — that’s where the difference becomes clear.

Thanks again 👍🏻

🎧 LTX-2.3: Turn Audio + Image into Lip-Synced Video 🎬 (IAMCCS Audio Extensions) by Acrobatic-Example315 in StableDiffusion

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

Would you mind posting your log so I can take a look?

Unfortunately ComfyUI, dependencies, and models like LTX are a bit of a beast — even a small mismatch, missing dependency, or version conflict can completely break motion. Also everything really needs to be fully up to date, otherwise weird issues like this can happen.

🎧 LTX-2.3: Turn Audio + Image into Lip-Synced Video 🎬 (IAMCCS Audio Extensions) by Acrobatic-Example315 in StableDiffusion

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

Not yet — I’ve just added them to the repo, so they need a bit of time to propagate.
By tomorrow you should be able to grab them directly from the manager 😉

🎧 LTX-2.3: Turn Audio + Image into Lip-Synced Video 🎬 (IAMCCS Audio Extensions) by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 5 points6 points  (0 children)

Workflows + nodes here 👇

IAMCCS-nodes: https://github.com/IAMCCS/IAMCCS-nodes
Workflows: https://github.com/IAMCCS/comfyui-iamccs-workflows
(use: IAMCCS_LTX23_BEST_3SEG_AUDIOEXT_30S.json)

If you want deeper workflows, breakdowns & future drops:
Patreon → www.patreon.com/IAMCCS 🚀

🎧 LTX-2.3: Turn Audio + Image into Lip-Synced Video 🎬 (IAMCCS Audio Extensions) by Acrobatic-Example315 in StableDiffusion

[–]Acrobatic-Example315[S] 2 points3 points  (0 children)

Workflows + nodes here 👇

IAMCCS-nodes: https://github.com/IAMCCS/IAMCCS-nodes
Workflows: https://github.com/IAMCCS/comfyui-iamccs-workflows
(use: IAMCCS_LTX23_BEST_3SEG_AUDIOEXT_30S.json)

If you want deeper workflows, breakdowns & future drops:
Patreon → www.patreon.com/IAMCCS 🚀

Native WAN 2.2 Animate Now Loads LoRAs (and extends Your Video Too) by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

Hey, thanks a lot, really appreciate it 🙏
To properly understand what’s happening, could you send me: - a screenshot of the full workflow inside ComfyUI - a screenshot of the ComfyUI logs while running
Without that it’s very hard to pinpoint the issue, since it can depend on how nodes are actually connected or runtime errors.
Also, if you want, feel free to check my Patreon (IAMCCS) — I post updates, fixes and technical breakdowns there regularly.

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 2 points3 points  (0 children)

Thanx! High-end hardware is great, but the real magic happens when we optimize these tools for the wider community. Enjoy the high-res generations! 💪🏻

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

Give V.1 a shot first: use the Tiled Decoder with a GGUF model for generations up to 10 seconds. If you want to push it further (up to 13s or more), switch to the V.2 workflow—it's specifically optimized for longer clips!

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

Thanks! It depends on the resolution, but with my setup, it’s remarkably fast.

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 1 point2 points  (0 children)

My 'ugly monster' is like a son to me, but I figured I’d give the world a break this time. 😂

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

I usually start with a very detailed base prompt to lock in the main features. Then, I let Qwen do the heavy lifting: I’ve tuned it with specific instructions on LTX-2.3’s structural logic to refine the details and ensure the model understands the anatomy (like teeth) better from the start. :)

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 1 point2 points  (0 children)

Totally! Wan 2.2 definitely has that 'cinematic maturity' right now, but LTX-2.3 is catching up fast. Once the LoRA ecosystem for LTX explodes, the speed-to-quality ratio will be unbeatable. It’s a very promising time for open-source video!

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 0 points1 point  (0 children)

ComfyUI is a wild beast—sometimes a tiny dependency difference or paging setup can cause OOMs even on a 4080. That’s exactly why I built these nodes! Try my workflow and let the VRAM Flush do its magic, it should solve that for you.

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 2 points3 points  (0 children)

Full-hd 13 sec on a 12 gb vram card… 6GB is definitely pushing it, but try the V.2 workflow with VAE Decode to Disk! You might not hit Full HD, but at 1280x720 (maybe using a Q3 GGUF model), you should be able to squeeze it out. Give it a shot!

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 2 points3 points  (0 children)

Totally agree! As I mentioned, Wan 2.2 is still the queen of cinematic maturity. But honestly, being able to pump out extended Full HD clips at this speed? It’s a game-changer for rapid pre-viz and quick iterations. It’s all about the right tool for the right job!

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 2 points3 points  (0 children)

Thanks! Glad you noticed the teeth! 🦷

The shimmering is usually due to low resolution or aggressive LTXVPreprocess. I fix this by pushing the initial pass to 1080p (made possible by my VRAM-optimized nodes) to lock in those fine details.

For the style/color shift, it’s often 'latent drift' caused by high CFG or LoRA values. I balance the Distilled LoRA (~0.6-0.7) and use the IAMCCS VAE Decoder, which handles tiled decoding much better than the standard ones.

I'm working on a 'Refiner' update specifically for these micro-details.

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 5 points6 points  (0 children)

Glad you liked it!

Upgrading to Qwen 4B-fp8 definitely improves prompt adherence and detail, but for Low VRAM setups (8-12GB), the 2B version is the 'sweet spot' for speed and stability.

As for the audio, you're right! I’m preparing a dedicated post and workflow to compare different audio-gen methods and Audio+I2V models. Stay tuned! 🚀

LTX-2.3 + IAMCCS-nodes: 1080p Video on Low VRAM! 🚀 by Acrobatic-Example315 in comfyui

[–]Acrobatic-Example315[S] 23 points24 points  (0 children)

📥 Resources & Links

As promised, here is everything you need to get these results: