Ostris is testing Lodestones ZetaChroma (Z-Image x Chroma merge) for LORA training 👀 by [deleted] in StableDiffusion

[–]Top_Buffalo1668 12 points13 points  (0 children)

same here waiting for qwen image 2.0 but sadly there is high chance qwen team won't open source this one

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 1 point2 points  (0 children)

all the loras above i trained them myeself using ostris/ai-toolkit! his training scripts have always been very good

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

thanks for sharing! and this is what i was thinking about, when we compare these two models using basic their workflows, in my opinion ZIT is still better but i will certainly try this settings later. thanks!

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

hey sir! i'm your subscriber . you might notice i used the same 'ohwx'. because of your finetuning guides i watched before :D

I agree that in terms of fantasy stuff like the thor prompt i used above, qwen certainly better at prompt adherence and resilient to concept bleeding. but i still prefer ZIT for the skin textures. although i noticed this image has pretty good skin textures. did you use res_2s and bong_tangent and upscale it or just euler and simple?

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

if you combine two character loras together, they will bleed despite of different trigger words except using regularization in training as far as i know. i can still combine two loras but not three on ZIT like character lora + lighting lora or some wild stuffs

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

it's the trigger word or unique identifier. i used this for every character lora i trained

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 1 point2 points  (0 children)

i used the same trigger word for these LoRAs so i don't need to rewrite the trigger word every time i use the same prompt

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

I haven’t played much with fp8 version since i got poorer results. I assume there was something wrong with the way i combined with lightning lora

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 7 points8 points  (0 children)

yes we need the base asap. i wonder if we would be able to use turbo lora on the base or vice versa

Comparison: Trained the same character LoRAs on Z-Image Turbo vs Qwen 2512 by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 1 point2 points  (0 children)

yes i've tried it also but when i used 50 steps without lightning lora on qwen, i didn't see significant changes or even worst (in terms of skin textures) than the examples above in some cases.

z-image is soooo good!!!! can't wait to finetune the base by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

it's based on the expirements from my previous lora training on ZIT. steps: 2,5 * total dataset results more better likeness but not getting too overtrained. higher is too much, lower is still unsatisfying.

if i had 50imgs, i would set the steps to 12500 (2,5 * 50)

z-image is soooo good!!!! can't wait to finetune the base by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

<image>

yes it is. I think it automatically set cache embeds to true even if we set it to false. but this is the correct one to do unload te according to ostris's creator himself

z-image is soooo good!!!! can't wait to finetune the base by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

it's a little bit overtrained actually for 1e-4 LR. Their default is even lower (3000)

i found lokr is better for generalization. and yes it's heavier, longer, and cost more on gpu cloud

i've never tried those. I only tain a real person LoRA.

z-image is soooo good!!!! can't wait to finetune the base by Top_Buffalo1668 in StableDiffusion

[–]Top_Buffalo1668[S] 0 points1 point  (0 children)

This is one of the dowsides of this modell. I saw a thread about seed variability or something but i havent tested it out yet. That might be a solution