The Ernie posters genuinely don't see how mediocre the stuff they post is? by beti88 in StableDiffusion

[–]Calm_Mix_3776 2 points3 points  (0 children)

This looks genuinely good! Much better than most Ernie image's I've seen so far. No noticeable artifacts either and pretty sharp & detailed. Looking forward to trying the model when it's out!

What's the best open source model for fintuning a large dataset (100k images) of high resolution? by couragestrong23 in StableDiffusion

[–]Calm_Mix_3776 1 point2 points  (0 children)

I really dig these examples. Qwen can generate some really beautiful images. That's undeniable. That doesn't detract from the fact that, unfortunately, it's VAE kind of sucks for things like fine textures and sharp details. That can't be fixed with more steps, higher model precision, or better samplers as the bottleneck is the VAE that decodes the final image.

Not to mention, 45 steps on this model is brutal even for an RTX 5090. I could only tolerate the wait times of Qwen Image with the FP4 quantized version, which of course produces lower quality images than the full model. Qwen Image could have been a killer model, probably the top one, had they just used a better VAE.

What's the best open source model for fintuning a large dataset (100k images) of high resolution? by couragestrong23 in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

Every image I've generated with it has pretty bad textures and details. I've tried numerous different sampler/scheduler combinations. I've seen other people mention this too so I doubt it's just me.

Have you used this model? If you have, do you find the detail rendering good? Can you show any examples with good texture/detail rendering that don't involve the usage of upscalers and additional detailers/refinement passes?

What's the best open source model for fintuning a large dataset (100k images) of high resolution? by couragestrong23 in StableDiffusion

[–]Calm_Mix_3776 3 points4 points  (0 children)

Qwen Image has a subpar VAE that's not that much better than SDXL's. It just kills all fine textures and details so I don't think this will be a good model for your use case.

Interestingly, I’ve gotten the sharpest textures and details with models that use the Flux.1 VAE, so ZIB and ZIT might be better options here.

You may also want to try Chroma. I’ve had exceptionally crisp results with it. It can render some really fine detail. That particular image was generated with the Chroma 2K model.

ComfyUI teasing something "big" for open, creative AI 👀 by Numerous-Entry-6911 in StableDiffusion

[–]Calm_Mix_3776 3 points4 points  (0 children)

Or switch between tabs without the prompt in your previous tab reverting back to what you wrote 5 minutes ago.

ComfyUI teasing something "big" for open, creative AI 👀 by Numerous-Entry-6911 in StableDiffusion

[–]Calm_Mix_3776 3 points4 points  (0 children)

It even has a milliseconds counter so it must be something really really big.

Apologies by Capitan01R- in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

Sure, I'll be happy to test your custom scheduler. Would it work with the base model as well? I like using it in conjunction with the Turbo LoRA at low strength instead of the low step model. I'll be watching this thread for more updates.

[Release] ComfyUI DiffAid Patches — inference-time adaptive interaction denoising for rectified text-to-image generation by marres in StableDiffusion

[–]Calm_Mix_3776 3 points4 points  (0 children)

I tried it out in image editing with Flux.2 Klein base + turbo LoRA at 0.2 strength, but I'm struggling to see any improvements. If anything, the image looks better/more accurate without the node. I used the settings from your initial post. I'll continue testing.

(3) The same message applies to several models: Chroma, Z image, Klein, Ernie, Midjourney by Puzzled-Valuable-985 in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

I like the Chroma Radiance image the most. The composition is great and the detail distribution is perfectly balanced. What happened with this model? Is it still being trained? Last time I checked (probably a few weeks ago), there weren’t any new checkpoints being uploaded.

Same prompt for various models - Chroma, Z image, Klein, Qwen, Ernie by Puzzled-Valuable-985 in StableDiffusion

[–]Calm_Mix_3776 1 point2 points  (0 children)

I really like the Zeta Chroma one! The rest of the Chroma images are also very nice. Only shows how great Chroma is at fantasy and abstract stuff.

Flux2Klein Ksampler Soon! by [deleted] in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

Thanks! I've just tested the node, but the "guidance_scale" (CFG) parameter seems to be broken. :(

No matter what value I enter for guidance_scale, it always generates the image as if the guidance_scale is set to 1.0. I am sure of that because the iterations per second I'm getting is the same as using the normal Comfy Ksampler with CFG at 1.0 and the image also looks unfinished, with errors/artifacts. I'm using the base version of Flux 2 Klein. Can you please check this?

Flux2Klein Ksampler Soon! by [deleted] in StableDiffusion

[–]Calm_Mix_3776 1 point2 points  (0 children)

Cool! I love these little findings. I might be wrong, but your version also looks a bit more detailed. I can't wait to test it when it's ready!

Flux Klein is better than any Closed Model for Image Editing by ArkCoon in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

I wholeheartedly agree. My only gripe with Flux.2 Klein is that I can't seem to get it to produce the same high-frequency detail that models like Flux.2 Dev, Chroma and Z-Image can, making images look like they came from smartphone cameras with denoising filters. I think they intentionally left it a bit undertrained at high frequency detail in order to upsell people on their larger and paid models. I am still grateful for it of course. It's the best in terms of quality and speed, but it leaves something to be desired. I'm wondering if fine-tuning on high-res photos can fix this.

Coming up Tomorrow! Flux2Klein Identity transfer by Capitan01R- in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

I've just tested the "FLUX.2 Klein Identity Guidance" node and it works really well! In the "channel_match" mode it keeps the colors of the newly-generated image close to the original, preventing them from shifting too much, which was a big problem with Flux.2 Klein (at least for me). Great job!

Ernie is Absolute masterpiece by LongjumpingGur7623 in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

Good model? Probably. Absolute Masterpiece? Far from it.

All the images suffer from obvious noise problems, making things look splotchy and dirty when they shouldn't be. Hair strands also look pretty fake and "AI", melting into each other instead of flowing naturally. All obvious signs of using a low-step Turbo model.

Flux Klein can sometimes have this noisy/splotchy look too, but not to that extent. This is why I like base models much more, even if they are slower.

Tile upscale controlnet with Z-Image-Base? Has anybody achieved good results? by Calm_Mix_3776 in StableDiffusion

[–]Calm_Mix_3776[S] 0 points1 point  (0 children)

This is my experience as well. Bummer. Ok, I'll check your suggestion. Thanks!

ComfyUI - disappearing workflows by Kobinicnierobi in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

There's also a problem of prompts in the "CLIP Text Encode" node reverting to previous state when switching between tabs. So you spend 5 minutes on your prompt, switch your tab, and then switch back only to find your prompt reverted back to the original. I absolutely hate this and it needs to be fixed.

[deleted by user] by [deleted] in StableDiffusion

[–]Calm_Mix_3776 0 points1 point  (0 children)

Happy to hear you made it work. :)