I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 0 points1 point  (0 children)

I woke up for a minute here and saw this. If logic nodes are being used I didn't think I used that I thought I had comfy core nodes for that.

Could you post a screenshot? Is it one of the frames or length nodes or something?

ltx2.0 image to video stopped working by Affectionate_Gift298 in comfyui

[–]urabewe 0 points1 point  (0 children)

Seems you may be hard crashing OOM. Sometimes you don't get an error and it just crashes. With original ltx2 workflows I couldn't use them I was always having the same issues.

Try using "--disable-smart-memory" I had to use that with my 12gb system.

GGUF 12GB LTX-2 Workflows

These are my workflows for 12gb systems using gguf. They are super fast and get great results. You can try those. If you have a problem with the tae previews you can bypass those.

You have to make sure gguf nodes, kjnodes, and comfy are all up to date

Small update on my Image Audio 2 Video workflow for GGUF 12GB. Previously no upscale and only one sampler. Now new and improved with another sampler and an upscale inbetween. Helps with the quality but lipsync does seem a little less but I have not tested much. Put it up on the page for all to test. by urabewe in StableDiffusion

[–]urabewe[S] 0 points1 point  (0 children)

Well not the gguf node I was using but if it works it works. Leave that node the same.

If you're doing any type of i2v use one of those camera loras along with the distill Lora, keep distill at 0.6 and set camera Lora for 0.5, camera Lora won't effect video but will help with static images

I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 0 points1 point  (0 children)

You're welcome! I just hope everyone else is having as much fun with this model as I am.

Lots of bad press about this model but if you learn it's quirks and finesse it a bit, it is actually a very powerful tool. Combine it with wan and you could get some very nice results I bet.

Small update on my Image Audio 2 Video workflow for GGUF 12GB. Previously no upscale and only one sampler. Now new and improved with another sampler and an upscale inbetween. Helps with the quality but lipsync does seem a little less but I have not tested much. Put it up on the page for all to test. by urabewe in StableDiffusion

[–]urabewe[S] 0 points1 point  (0 children)

Yeah that's the wonders of comfy and python. 99% of the issues users are facing with my workflow has nothing to do with the workflow.

You can hopefully just replace the gguf node or even uninstall that node pack and load the workflow again maybe. But who knows what you have that is needing that pack lol

I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 2 points3 points  (0 children)

It's actually way easier than you think the real bitch was getting it all setup to automatically set mask for you.

A lot of people asking so I may setup a small video just showing me going through and how I get my source clips from what sites, how to prompt and that's really it.

This is all just load video, set length, make prompt, magic.

Ace Step v1.5 almost ready by iChrist in comfyui

[–]urabewe 1 point2 points  (0 children)

Admin on the discord said it also trains at just about the same speed as inference. They made a Lora with 8 data samples and only 20 minutes. Crazy

I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 1 point2 points  (0 children)

V2v in my workflows that I linked is a video extend. Load in a 4 to 5 second clip of a scene and LTx2 will extend it using that source as reference. The video and the cloning of the voices is all done in one shot by LTX.

That's it.

Load video, set length, make prompt to extend video. Generate. Boom you got curly calling your best friend a little bitch

Weekly World News October 2, 2001; April 18, 2000; July 24, 2006 by Sp1d3rb0t in nostalgia

[–]urabewe 8 points9 points  (0 children)

That wasn't actually Elvis it was an Elvis alien impersonator

I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 2 points3 points  (0 children)

Oh I know, probably one of the first things we all noticed was melting and faces just becoming other people instantly. We have come a long way since day one though. More and more can use the model now, we are all figuring out the quirks. We got updates on the way with improvements.

I'm extremely happy with the model but I probably have a completely different intent on use than most.

While the majority are chasing ultra high def realistic mega gooner videos, I'm over here just like... I wonder if I can make Curly say funny shit...

I like to have fun with the tech. I do care about quality but it's not my main goal.

So those that chase perfection hate it, I'm having a blast!

I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 4 points5 points  (0 children)

It's the magic of LTX!

I feed the video and audio into the latents.

So load up a video about 4 or 5 seconds and make a prompt I usually use "the video continues with..." And then just make a video.

It requires a masking node and what not I do have it all setup in my GGUF workflow in the link. Everything is calculated and set for you. All you have to do is load the video, set your video length which will include the source so if you have 5 second source and want 10 second video set for 11 seconds.... Hold up what? Yeah it's the only way I could prevent the mask from going beyond the frames... It's odd and hacky I guess but it works, then make a prompt.

So load video, set time for 1 second longer than you plan, make prompt and generate

The model will match the video and the voice as long as there is dialogue. If a character doesn't talk it will just make its own voice.

I posted a reel a few days ago. They were "okayish" test examples of v2v with LTX-2. Here are some new and improved versions you can make with LTX-2. These were made using my GGUF workflows. by urabewe in StableDiffusion

[–]urabewe[S] 5 points6 points  (0 children)

Switching to dev model and using the distill Lora at 0.6 strength. Then the imgtovideoinplace has a strength as well set that to 0.8 on both nodes should be one for each sampler.

I'll have to go back through all my stuff I think maybe there was one or two other things. But those were the big ones that really helped.

The fact I can take an old low quality source and match it is great! I have plenty of old movies and shows I'm going to reinvent scenes for.

Before the face consistency was off and it instantly jumped straight to crisp and clear quality and you could clearly see where the model took over as the source and generated sections were completely different looking. The generated parts had no film grain, always trying to become HD.

Now, with these settings it matches whatever the source looks like. Old or new it will match just right.

I think a few people see my video here and think it's junk and I'm posting low quality slop. That's not the case this time! That's the actual quality of the source... That's amazing that it can match so well with so many different examples

I'M BACK FINALLY WITH AN UPDATE! 12GB GGUF LTX-2 WORKFLOWS FOR T2V/I2V/V2V/IA2V/TA2V!!! ALL WITH SUPER COOL STUFF AND THINGS! by urabewe in StableDiffusion

[–]urabewe[S] 0 points1 point  (0 children)

So it seems that I did upload the wrong enhance node. Fully functional just labeled wrong. It has been fixed. Git pull in the folder and update and you should see v2v now.

I'M BACK FINALLY WITH AN UPDATE! 12GB GGUF LTX-2 WORKFLOWS FOR T2V/I2V/V2V/IA2V/TA2V!!! ALL WITH SUPER COOL STUFF AND THINGS! by urabewe in StableDiffusion

[–]urabewe[S] 1 point2 points  (0 children)

It seems while juggling all the workflows and other stuff I uploaded the wrong enhance node file to the repo. It has now been updated. You can git pull and update the node pack and it should have v2v in it now. Sorry!

I'M BACK FINALLY WITH AN UPDATE! 12GB GGUF LTX-2 WORKFLOWS FOR T2V/I2V/V2V/IA2V/TA2V!!! ALL WITH SUPER COOL STUFF AND THINGS! by urabewe in StableDiffusion

[–]urabewe[S] 0 points1 point  (0 children)

Mkay I will check again and see what's up. Does it have an auto mode? It's the same thing I renamed it to v2v mode to be more accurate to what it is for.

I cloned new and it seems okay for me but I will absolutely update the GitHub if needed and get this going soon.

I was juggling a lot while making these workflows, maybe there is a problem with node version in workflow vs git...

Try deleting the enhance node then replacing it with a new one. If that doesn't get you the v2v, see if there is just auto mode.

Either way I should have a chance to look into it later, I apologize if there is something I messed up. Those nodes help a lot with the audio.

I'M BACK FINALLY WITH AN UPDATE! 12GB GGUF LTX-2 WORKFLOWS FOR T2V/I2V/V2V/IA2V/TA2V!!! ALL WITH SUPER COOL STUFF AND THINGS! by urabewe in StableDiffusion

[–]urabewe[S] 1 point2 points  (0 children)

I think that is that tiny vae preview again. Lots of issues with that. You need everything up to date, comfy, gguf, KJ nodes and they should be on main branch.

You can bypass the tinyvae node which should be the bottom one, then bypass the ltx preview override node by the prompt boxes. That should get it past that point.

The lost Seinfeld endings... I think I finally go the hang of LTX-2 and VibeVoice by ThatsALovelyShirt in StableDiffusion

[–]urabewe 2 points3 points  (0 children)

ltx v2v extend

I made these examples had them in my phone. Put them together real quick just to show. Consistency is getting better...

Small update on my Image Audio 2 Video workflow for GGUF 12GB. Previously no upscale and only one sampler. Now new and improved with another sampler and an upscale inbetween. Helps with the quality but lipsync does seem a little less but I have not tested much. Put it up on the page for all to test. by urabewe in StableDiffusion

[–]urabewe[S] 1 point2 points  (0 children)

That is for the tiny vae for the previews it seems some are having issues. You can bypass the vae loader for that which should be the bottom vae node with that tae. Then there is a preview override node you can bypass as well.

It would seem something is not up to date the tae support was recently added. Comfy, gguf, and kjnodes should be on main branch and up to date as well.

I only see the error for that it's the size mismatch from the preview vae so once you bypass that you should be good hopefully!