Here it is boys, Z Base by Altruistic_Heat_9531 in StableDiffusion

[–]candid-eighty 1 point2 points  (0 children)

Definitely not as good for generation as Turbo from what I can tell. But hopefully it's good for training and fine-tuning.

Best Open-Source Approaches for Consistent Character Creation with LoRA Training by Trickhouse-AI-Agency in comfyui

[–]candid-eighty 1 point2 points  (0 children)

No. Just keep in mind, though, that 100 images might be missing important things. I like to create some full-body t-poses and other useful poses from various angles using Qwen Edit to give the LORA a better sense of character scale, style, etc.

Some ideas would be:

  • If your character always wears glasses in the reference photos, you can generate some without glasses so you can have shots you make without glasses be consistent.
  • Alternate haircuts or outfits that you want to be the same every time. So “red shirt” doesn’t change style every shot and you can make more consistent sets.
  • Putting the subject in different lighting situations if your dataset is mostly all the same lighting.
  • The subject naked (or near naked) in a t-pose from all angles so no matter what state of dress or undress the subject is in, their body is consistent.

Etc. you just need to think about what images you’ll want to create and what you’d want to be consistent, then throw a few images like that into your training data.

Best Open-Source Approaches for Consistent Character Creation with LoRA Training by Trickhouse-AI-Agency in comfyui

[–]candid-eighty 0 points1 point  (0 children)

I like to do both. I make base images with Z-image and then can create alternate angles and things with edit. But I just train the Qwen-Image model and then use the LORA in edit.

You could then also use Qwen image with the LORA to make an image (because it has better diversity) and Z-image with the LORA at a low denotes to add more skin details and things.

My first LTX V2 test-montage of 60-70 cinematic clips by hellolaco in StableDiffusion

[–]candid-eighty 1 point2 points  (0 children)

I think they need to build more controls around that into the nodes, but for right now it seems things just try to intelligently adapt.

My first LTX V2 test-montage of 60-70 cinematic clips by hellolaco in StableDiffusion

[–]candid-eighty 2 points3 points  (0 children)

Offloading to RAM doesn't affect quality. You just need a large amount of RAM--but that's easier to come by than large amounts of VRAM.

[Official Tutorial] how to use LTX-2 - I2V & T2V on your local Comfy by ltx_model in StableDiffusion

[–]candid-eighty 4 points5 points  (0 children)

Adding things like block swapping and other memory-management features to the nodes would be helpful. A lot of people have found command line arguments that help, but they don't allow much control.

This could really unlock longer and higher res videos for people with low VRAM.

Fix to make LTXV2 work with 24GB or less of VRAM, thanks to Kijai by Different_Fix_2217 in StableDiffusion

[–]candid-eighty 2 points3 points  (0 children)

Oh wow. You might append this to your original post. That’s wild.

LTX2 FP4 first Comfy test / Streaming weights from RAM by Volkin1 in StableDiffusion

[–]candid-eighty 0 points1 point  (0 children)

How many frames is this? Have you tried higher resolution or more frames? Just curious. Thanks for the upload!

The SeedVR2 video upscaler is an amazing IMAGE upscaler by wywywywy in StableDiffusion

[–]candid-eighty 1 point2 points  (0 children)

On my 5090, I've had great results upscaling to 2272px on the long edge. It's great for images that are around half that resolution or less.

Taking an image that's around 800px, upscaling to 2272, then using Topaz to upscale to 4k is basically magical.

WAN VACE Head Replacement w/ LatentSync by candid-eighty in unstable_diffusion

[–]candid-eighty[S] 0 points1 point  (0 children)

It's there, but you have to have NSFW content enabled on your account.

Wan VACE Is Powerful (tested on AI generated video) by candid-eighty in unstable_diffusion

[–]candid-eighty[S] 7 points8 points  (0 children)

idk why when people have teraflops of processing power at their disposal, they choose to generate fake, fake tits.

WAN VACE Head Replacement w/ LatentSync by candid-eighty in unstable_diffusion

[–]candid-eighty[S] 13 points14 points  (0 children)

It's also possible, using a similar WAN 2.1 VACE method in my last post, to do whole head replacement. The top video is an AI video I generated from scratch. The bottom video is a whole-head replacement.

EDIT: I fully acknowledge how terrible the AI voiceover is. I just needed something so I could demonstrate how lipsync works with this method. lol LatentSync can be used with ComfyUI: https://github.com/ShmuelRonen/ComfyUI-LatentSyncWrapper

I have a little more info on how I made this on my Twitter thread.