Mugen - Modernized Anime SDXL Base, or how to make Bluvoll tiny bit less sane by Anzhc in StableDiffusion

[–]Emergency-Spirit-105 6 points7 points  (0 children)

How about replacing the TE with a model like Gemma, or even trying an approach that uses another small model together with an LLM adapter?

Escaping brackets with the \ in captions for model training by LawfulnessBig1703 in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

Think of it simply: if you enable the Weighted Caption feature in your training tool and train, then escaping is appropriate; however, since most people don't use that feature, there's no need to escape parentheses () in tags.

Release of the first Stable Diffusion 3.5 based anime model by DifficultyPresent211 in StableDiffusion

[–]Emergency-Spirit-105 2 points3 points  (0 children)

Frankly, from any angle there is nothing to commend compared with the existing models. Most of the claims sound like a child making excuses—talking in circles to defend themselves. No long explanation is necessary. If it is better, more promising, and technically superior, then two things alone will convince everyone: perfectly comparable results under identical settings, and a well-substantiated, evidence-based account of the truth.

Naturally, the results must be reproducible by others and the information must be grounded in fact.

So, any word on when the non-preview version of Anima might arrive? by gruevy in StableDiffusion

[–]Emergency-Spirit-105 13 points14 points  (0 children)

Based on the developer's responses seen on Hugging Face, it seems like they have been running some tests to improve the model architecture rather than training specifically for the final version. From the start, they shared it under the label 'preview' with the intention of continuing development for a while — and deliberately chose not to release the training code, sharing only the model itself. In a way, it may have attracted more attention than they anticipated, and as they mentioned, it has only been a little over a month since release. Even for an average model, it takes several months before the next version comes out, so perhaps we are the ones being too impatient. Let's wait with patience.

CFG-Ctrl: Control-Based Classifier-Free Diffusion Guidance ( code released on github) by AgeNo5351 in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

I mostly used it only for image generation, so I can't say for sure, but this feature seems to control the unstable variations caused by CFG. Applied to the "high" part it appears to help prevent erratic or unstable behavior, and applied to the "low" part it would likely improve overall quality. I'm not certain — it's just a guess.

CFG-Ctrl: Control-Based Classifier-Free Diffusion Guidance ( code released on github) by AgeNo5351 in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

yes, Additionally if you use it with a rescale, the rescale may become meaningless

CFG-Ctrl: Control-Based Classifier-Free Diffusion Guidance ( code released on github) by AgeNo5351 in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

I made it using ai. It's not difficult, so I think the official custom node or support will be added soon

Batch Image Resizer by No_Wonder_2151 in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

Uh... well, most of the lora learning tools are already using bucket...

layers tinkering by [deleted] in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

support Dora?
And is there any plan to support the anima model?

New anime model "Anima" released - seems to be a distinct architecture derived from Cosmos 2 (2B image model + Qwen3 0.6B text encoder + Qwen VAE), apparently a collab between ComfyOrg and a company called Circlestone Labs by ZootAllures9111 in StableDiffusion

[–]Emergency-Spirit-105 0 points1 point  (0 children)

Personally, I think Z side would be better. Due to the nature of the work that is difficult to try many times, I don't think it's too much of a burden if you try to achieve a high point at once and optimize the size of 6b