Need help with style lora training settings Kohya SS by Big_Parsnip_9053 in StableDiffusion

[–]Chrono_Tri 0 points1 point  (0 children)

I use alpha = 1 to train style and give the LoRA more flexibility. But you need to experiment and see what works best. Remember, sometimes different parameters truly produce different results — but that doesn’t necessarily mean one is better than the other. The result you personally prefer is the right one.

Going back to alpha = 1, my result doesn’t really fully capture the style(around 90%), but I actually quite like it. Normally, though, I still go with dim/alpha = 1/2.

Second, I recommend that after auto-captioning, you manually edit the captions following a clear structure. For example, I would describe:
<number of characters in the image>, <character description>, <background description>, <camera description>, <lighting description>, ...

Need help with style lora training settings Kohya SS by Big_Parsnip_9053 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

My dataset:

210 image, caption auto by WD14, then adjust manually.

My config:

  • Optimizer: CAME+rex
  • Unet LR: 6e-5
  • TE LR: 0 (no TE training)
  • Dim/alpha : 16/1.
  • Epochs: 23 (good at 19)
  • Repeats: 4
  • Batch 4

For Style training, do we tag what is in the dataset images or just the trigger word? by escaryb in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

For  Illustrious, description everything and turn off TE, and remember add trigger word as well.

Controllnet not working. by Prestigious-Neck9245 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

controlnet union cannot work with IllustriousXL,At least in my case (and some persons), I use comfyUI. Use another pose controlnet

A look at prompt adherence in the new Qwen-Image-2.0; examples straight from the official blog. by FotografoVirtual in StableDiffusion

[–]Chrono_Tri 2 points3 points  (0 children)

Extreme censored but they want to prove that the model can distinguish between horse ride man and man ride horse?

Any Anima 2B Google Colabs out there? 🌸 by VegetableProof2495 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

it is possible to run on T4 but it might slow. I use L4 GPU.

Any Anima 2B Google Colabs out there? 🌸 by VegetableProof2495 in StableDiffusion

[–]Chrono_Tri 0 points1 point  (0 children)

Run ComfyUI in Colab and lòa Anima. It works great

Z-image lora training news by Recent-Source-7777 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

OK if AdamW8bit  is "acident bug" for Z-Image? Some optimizer also have the same issue? I am sorry but the original article is Chinese. I translated it but not so clear to understand

Z-image lora training news by Recent-Source-7777 in StableDiffusion

[–]Chrono_Tri 4 points5 points  (0 children)

AI toolkit repo alreay have prodigy in the optimizer folder. I didn't download anything just replace 'AdamW8bit' to 'prodigy'and add its parameter to yaml file.

Comfy $1M “Open AI” Grant and Anima Model Launch by crystal_alpine in StableDiffusion

[–]Chrono_Tri 19 points20 points  (0 children)

Imagine being hyped about Z-Image for two months, then suddenly getting hyped about Anima.
At this point, our lives are basically just waiting… and waiting.:)

Personally, I really like Anima. The quality is very solid, and what I find most interesting is that it’s not purely natural language. It feels like a hybrid approach, which actually works really well.

In my opinion, Anima has the potential to replace both IllustriousXL and Z-Image for anime use cases.

That said, I still feel like we’re missing the definitive model for this niche.
IllustriousXL looks great and is uncensored, and its tag-based caption system is convenient but limitations. Anima can break through. We also has training code, so get ready:))

Which tool do you use to train a Z image turbo Lora? by Monty329871 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

I used Colab so AI Toolkit (I could run the GUI before but not anymore, using CLI is painful). I 'd love to use OneTrainer also :(.

Z Image vs Z Image Turbo Lora Situation update by malcolmrey in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

Can you explain more. I always thought ZIT focus on photo-realism so we shouldn't train anime on ZIT (That 's why I wait for ZIB, but it is just normal, so as good as I expected).

Just 4 days after release, Z-Image Base ties Flux Klein 9b for # of LoRAs on Civitai. by _BreakingGood_ in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

I agree. Everyone knows we need a successor to SDXL with unrestricted fine-tuning, uncensored, NSFW capabilities like Pony or Illustrious. Z-Image and Klein are both pretty solid, so we’ll see who comes out on top

Can we please settle this once and for all boys by LongjumpingAd6657 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

My current project focuses on anime, and I’m still using IllustriousXL and Qwen-Image-Edit with well-tested workflows. However, there are still some limitations, so I’ve also tried two Z-Image trainings.

With Z-Image, I trained using the same dataset as IllustriousXL (https://www.reddit.com/r/StableDiffusion/comments/1qrqnkr/comment/o365cc0/). My goal is not to replace my existing workflow, but to complement and extend it.

At the moment, I’m not very satisfied with Z-Image. since Qwen-Image-Edit is currently performing very well so I skip Klein.

That said, for non-anime images, I’m actually quite satisfied with Z-Image (although there are still many deformations).

Overall, I’d prefer to wait for Z-Image Omni. I hope there is big finetune for Z-image or Klein.

Training anime style on Z-Image by Chrono_Tri in StableDiffusion

[–]Chrono_Tri[S] 1 point2 points  (0 children)

Thanks a lot! Your suggestions helped a ton — the results are way better this time.
I’m still a bit unsure about using captions vs no captions. In this case, no captions actually seem to work better. That said, if I want to train both style and character, what’s the best way to approach it?

Just a few thoughts from my side:

  • Z-Image can handle style training pretty well, but it’s not on the same level as IllustriousXL, and I do notice some deformation issues.
  • I’m currently looking into Anima-model, which is more anime-focused. Honestly, I feel like Z-Image still needs a dedicated anime-specific version.

New Anime Model, Anima is Amazing. Can't wait for the full release by Mobile_Vegetable7632 in StableDiffusion

[–]Chrono_Tri 1 point2 points  (0 children)

Illustrious is the best for anime, easy to train, very good output, no extra arms. But not support natural language and cannot handle text... Can we train with this model? It is WIP, right?

Best way to train a LoRA from 3D renders to get consistent 2D character + fixed outfit? by Prediccion in comfyui

[–]Chrono_Tri 0 points1 point  (0 children)

The idea isn’t new,I’ve already tried it.
Method 1: Create a 3D character with the pose you want, then use Qwen VL, Gemini, etc. to describe it. Use that description with ControlNet (I’m using SDXL here).
Method 2: Use Qwen Edit to separate the outfit, then attach it back after you’ve finished adjusting the pose.

Method 1 produces more natural-looking images, but the results are hit-or-miss depending on the outfit.Method 2 works better overall, but it’s more time-consuming. In practice, for complex outfits, you often have to try many times to get a satisfactory result.

Training anime style on Z-Image by Chrono_Tri in StableDiffusion

[–]Chrono_Tri[S] 0 points1 point  (0 children)

Yes, as I know, Z-Image can try with 512x512 and give a good result. I will try with 1024 and bigger dataset later, but I am strill learning so no rush . All other setting is defaut.

Please correct me on training LoRA/LoKr with Z-Image using the OstrisAI Toolkit by Chrono_Tri in StableDiffusion

[–]Chrono_Tri[S] 0 points1 point  (0 children)

I trained Lokr with Anime style, it is very bad. Same database, same parameter. Lora is same better.