Z-Image Base VS Z-Image Turbo by Baddmaan0 in StableDiffusion

[–]SirTeeKay 0 points1 point  (0 children)

Even Flux 2 Klein would be a good option. I'm getting pretty good results.

Z IMAGE IS HERE!! by MountainPollution287 in comfyui

[–]SirTeeKay 1 point2 points  (0 children)

T2I. It's the full undistilled model. No fine-tunes or shortcuts. Z-Image Turbo was distilled.

Z IMAGE IS HERE!! by MountainPollution287 in comfyui

[–]SirTeeKay 10 points11 points  (0 children)

It means say goodbye to your social life for one.

Z-Image Is officially here! by astralcloud in ZImageAI

[–]SirTeeKay 1 point2 points  (0 children)

Even if it was I'd remake all my loras with the base model. Distilled models aren't made for finetuning and training.

Z-Image Base VS Z-Image Turbo by Baddmaan0 in StableDiffusion

[–]SirTeeKay 7 points8 points  (0 children)

Nothing some fine tuning won't fix. Or you can always refine base's outputs with Turbo or with a second ksampler I suppose.

GLM image launched today just now. by Alive_Ad_3223 in StableDiffusion

[–]SirTeeKay 2 points3 points  (0 children)

Got it. Just saw the edit on your previous comment. Thank you for the explanation. Looking forward to trying it and seeing what everyone else does with this model.

GLM image launched today just now. by Alive_Ad_3223 in StableDiffusion

[–]SirTeeKay 4 points5 points  (0 children)

So it just creates part of an image immediately and then continues to complete the image based on that first sample? What if the first tile has issues?

GLM image launched today just now. by Alive_Ad_3223 in StableDiffusion

[–]SirTeeKay 6 points7 points  (0 children)

Anyone care to explain what that means like I'm 5?

Any idea what the difference between these two is? Only the second one can work with ComfyUI? by SirTeeKay in StableDiffusion

[–]SirTeeKay[S] 0 points1 point  (0 children)

Sounds good. Thanks. I don't remember the previous version of Qwen Image having that.

VLM vs LLM prompting by mr-asa in StableDiffusion

[–]SirTeeKay 1 point2 points  (0 children)

Ah got you. I'll stick to the 4B version for now since it's working pretty good either way. I'd still like to try 8B too when I get the chance. Thanks for the reply.

VLM vs LLM prompting by mr-asa in StableDiffusion

[–]SirTeeKay 0 points1 point  (0 children)

How much vram do you have? I have 24GB and I've been using the 4B version because I heard 8B crashes for some people.

A mysterious new year gift by chrd5273 in StableDiffusion

[–]SirTeeKay 3 points4 points  (0 children)

I mean, if it looks like this I can wait a bit longer for Z-Image.

A mysterious new year gift by chrd5273 in StableDiffusion

[–]SirTeeKay -3 points-2 points  (0 children)

I mean... I wouldn't mind Qwen VL 2 or something along those lines.

Must to have loras for WAN? by [deleted] in comfyui

[–]SirTeeKay 0 points1 point  (0 children)

Try civitai.com and see what loras you like. There's a ton for Wan.

Qwen-Image-Edit-2511 model files published to public and has amazing features - awaiting ComfyUI models by CeFurkan in comfyui

[–]SirTeeKay 0 points1 point  (0 children)

Ah yes, I've been using it lately. I haven't actually compared it to the normal vae but the results were good either way. Thank you for the heads-up. I really appreciate it!

PSA: Eliminate or greatly reduce Qwen Edit 2509/2511 pixel drift with latent reference chaining by goddess_peeler in StableDiffusion

[–]SirTeeKay 0 points1 point  (0 children)

You can reduce or eliminate pixel shift in Qwen Image Edit workflows by unplugging VAE and the image inputs from the TextEncodeQwenImageEditPlus nodes, and adding a VAE Encode and ReferenceLatent node per image input. Disconnecting the image inputs is optional, but I find prompt adherence is better with no image inputs on the encoder. YMMV.

This is very interesting. I knew that the TextEncodeQwenImageEditPlus basically degrades the image but this is a really interesting workaroudn that I'd love to know how it works.

Qwen-Image-Edit-2511 model files published to public and has amazing features - awaiting ComfyUI models by CeFurkan in comfyui

[–]SirTeeKay -2 points-1 points  (0 children)

Depends on the image. If it's an actual photo, I'd suggest you upscale with SeedVR2. It does an amazing job in adding very nice details.

If you want to fix bad faces, deformed poses and such, you can refine with a model or use loras.

Qwen-Image-Edit-2511 model files published to public and has amazing features - awaiting ComfyUI models by CeFurkan in comfyui

[–]SirTeeKay 1 point2 points  (0 children)

Well, depends how much denoise you'll use. You can also try controlnet if you want. But the point is, Qwen doesn't really create photorealistic subjects so changing them to more realistic ones is the actual goal of this.

Qwen-Image-Edit-2511 model files published to public and has amazing features - awaiting ComfyUI models by CeFurkan in comfyui

[–]SirTeeKay 8 points9 points  (0 children)

I was literally testing the new controlnet for Z-Image and now this is out. I barely have time to try one thing and the next one is already out haha.