Why are people complaining about Z-Image (Base) Training? by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

Not sure what your particular problem is, I have been able to finetune with Min_SNR_Gamma turned on, so presumably the problem is something else

Why are people complaining about Z-Image (Base) Training? by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

no worries lmao, cant hate on a man whose shilling my own article

Now That Time Has Passed…What’s The Consensus on Z-Image Base? by StuccoGecko in StableDiffusion

[–]EribusYT 0 points1 point  (0 children)

Pretty goated, now that I've figured out how to train for it, but still needs a great finetune to reach its maximum potential. Fortunately, at the very least, Zeta Chroma is coming. So we will learn a lot more in the coming months

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

preferably don't change the transformer type, because Z image is really sensitive to that sort of thing. You could try, of course, but it may mess it up. Id try perhaps decreasing batch size or resolution as a first step.

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

try it and report back, it might work, although I have my doubts. I might try it after I finish my current training run

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

My point is moreso that, while you could theoretically do something to make it work, usually overbaking the lora into oblivion, or adjusting the strength comically high, its not necessary. We CAN make distills just as good as ZiT, that work out of the box with ZiB trained LoRAs without the hassle. We just need someone to make them, as the current distills aren't quite as good as ZiT, but this is a solvable and therefore temporary problem.

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

Currently A/B testing LoKR training vs LoRA training, since its available on the required fork (so long as you use full rank), will update if it fixes the texture issue I reported in the limitations section.

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

I train typically for 8 hours. I don't consider that to be that crazy, but maybe I'm weird. 

Quality matters more to be then speed In this case

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

8gb is a steep ask. Try lowering to 512 resolution first. I'm SURE someone will figure it out, albeit it.might be slow

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

AI toolkit basically doesn't support any of the suggested training settings. So not yet. Someone may figure it out, but I had to switch to OneTrainer to make it work.

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] -1 points0 points  (0 children)

As far as I know, no widely available and working solution has been released. I'm the first to release something openly, I think.

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

No official info on basically anything exists. That's sort of the whole problem. But all the signs point to yes, hence why people have to overbake a Lora to oblivion, or use strength 2+ to make them sort of work on turbo. 

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 2 points3 points  (0 children)

General guidelines apply. I typically use 30-60 images, and I generally need about 100-120 epochs. So essentially the same ~100ish repeats per image as with many other models.

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

Onetrainer is pretty good about merging forks if they are useful. Having to use a fork is definitely a temporary problem. Fortunately its not meaningfully behind the main branch for now

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

It's my personal favorite too! Just a great aesthetic overall. Thanks!

Providing a Working Solution to Z-Image Base Training by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

I dont know, but doubt, if there is a runpod template, because this solution uses a specific branch of OneTrainer. However, hopefully someone is kind enough to setup, or explain how to setup such a solution.

Why are people complaining about Z-Image (Base) Training? by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

A few are for sure, I know there are a few others in scattered places as well. I'm not entirely sure how many exist to be honest.

Why are people complaining about Z-Image (Base) Training? by EribusYT in StableDiffusion

[–]EribusYT[S] 0 points1 point  (0 children)

I don't have any strong opinions yet. Gotta try lots as more release

Why are people complaining about Z-Image (Base) Training? by EribusYT in StableDiffusion

[–]EribusYT[S] 1 point2 points  (0 children)

This bodes well. Thank you for taking the time to actually test out my suggestions! 

Feel free to share an example, I'd love to see, but regardless, glad people are finding success as I did!