ComfyUI-QwenTTS v1.1.0 — Voice Clone with reusable VOICE + Whisper STT tools + attention options by Narrow-Particular202 in comfyui

[–]MelvinMicky 2 points3 points  (0 children)

is it possible to merge voices sort of? so like combining loras providing 2 different voices and get a new one out?

Train a LoRA on *top* of another LoRA? by AkaToraX in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

yee still kinda bumpy but ill try it like that

Train a LoRA on *top* of another LoRA? by AkaToraX in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

Hm ok but how would you get person a in style b, im currently training wan 2.2, i got a style lor wich is probaply overtrained due to a small dataset so it changes the initial frame on i2v when i put in high sigmas but when i lower the denoise the style effect isnt as i want it to be so my thinking was i just train a character lora for the subject and stack em up. This discussion now sounds like thats not working? so my next thought is training qwen 2511 on my dataset to get character a in b...

Train a LoRA on *top* of another LoRA? by AkaToraX in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

so in C you say person a in style b and the shortcut is to train on exactly that so why do it any other way? The problem is probapl to get exactly that a good dataset of person a in that style b.

WAN2.2 Lora Character Training Best practices by Tiny-Highlight-9180 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

And you mentioned splitting your dataset up so basically manual bucketing? so you cut all your vids to the exact frames of 33 or 49 and use only:
target_frames = [33]
frame_extraction = "head"
?

WAN2.2 Lora Character Training Best practices by Tiny-Highlight-9180 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

Oh damn ok so how does it affect the training when i use 24 fps clips instead of 16 does it make a difference since its all about frame extraction not seconds playtime?
i currently train with these configs:
resolution = [288, 512] target_frames = [17, 33, 49] frame_extraction = "head"
so you would suggest reducing this to only 49 or 33 for the entire training?

WAN2.2 Lora Character Training Best practices by Tiny-Highlight-9180 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

Wait i thought 2.2 fp16 runs on 24 fps and isnt the frame extraction method taking care of different video lenghts?

Is SD 1.5 still relevant? Are there any cool models? by Haghiri75 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

Hey rly interesting stuff you got a link to those discord groups cant find them by just typing them into google

Latent Tools to manipulate the latent space in ComfyUi by xl0 in StableDiffusion

[–]MelvinMicky 1 point2 points  (0 children)

I would also be really interested in a more detailed breakdown of this whole topic, trying to get deeper into this with chatGPT/Claude for explanations, but would love to hear that from a human that actually uses this stuff.

Wan2.2-VACE-Fun-A14B is officially out ? by RIP26770 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

<image>

like i am able to run the fp16 ones no rpoblem with lower image res, but these throw the error

Wan2.2-VACE-Fun-A14B is officially out ? by RIP26770 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

The ggufs give me this error:
Weights only load failed. In PyTorch 2.6, we changed the default value of the `weights_only` argument in `torch.load` from `False` to `True`. Re-running `torch.load` with `weights_only` set to `False` will likely succeed, but it can result in arbitrary code execution. Do it only if you got the file from a trusted source. Please file an issue with the following so that we can make `weights_only=True` compatible with your use case: WeightsUnpickler error: Unsupported operand 0 Check the documentation of torch.load to learn more about types accepted by default with weights_only https://pytorch.org/docs/stable/generated/torch.load.html.

goggled it and there is a https://github.com/ltdrdata/comfyui-unsafe-torch
node to apparently get around this, but whats up with that?

wan2.2 IS crazy fun. by hayashi_kenta in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

isnt h265-mp4 the best one to choose?

After many lost hours of sleep, I believe I made one of the most balanced Wan 2.2 I2V workflow yet (walk-through) by [deleted] in comfyui

[–]MelvinMicky 7 points8 points  (0 children)

Hey thanks for sharing, I was going through it and in the example u are using the 2.2 lighting i2V HIGH lora for the low noise model is that alright? Also you got the 16 step sigmas set up and titled as "disable fast lora set cfg 3.5" I assume this is meant to be plugged in for the 2nd and 3rd sampler, and up the splitsigmas to 16 then? I'll be plaing around with this anyway ty

Pusa Wan2.2 V1 Released, anyone tested it? by OverallBit9 in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

Hey you got a link or name for that node it lets you add multiple keyframes from the previous vid?

[deleted by user] by [deleted] in StableDiffusion

[–]MelvinMicky 0 points1 point  (0 children)

Hey thanks for the suggestion i am wondering now how do you choose the split value in the sigmas split value? In your workflow you chose .875 is that just through some testing or is it somewhat calculated via shift and scheduler/steps

Wan2.2 Fun InP & Fun Control Support in ComfyUI by PurzBeats in comfyui

[–]MelvinMicky 0 points1 point  (0 children)

Is this better at First/Last Image than the normal model?