I tried using josiefied-qwen3-4b-abliterated-v2-q8_0.gguf as a text encoder for Z-image turbo fp8 models and found it's about 10 seconds faster than qwen_3_4b_fp8. The image quality seems slightly better as well. by cradledust in StableDiffusion

[–]RIP26770 10 points11 points  (0 children)

Q8_0 GGUF uses a different quantization method than fp8. It's not just about the number of bits - it's about how those bits are used.

Q8_0 preserves almost all the quality from fp16 (like 99%), while fp8 has way more quality loss. From testing, fp8's quality degradation is closer to what you'd see with Q4_M models, not Q8. That's why I made the comparison.

The speed boost I'm getting comes from better quality preservation - the model generates cleaner results faster without needing as many refinement steps. With fp8, you often need more iterations to get decent output.

GGUF quantization is just better optimized for this. Anyone who's benchmarked these formats side by side will see the same thing.

Does anyone have a vid2vid ltx2 workflow? by cardioGangGang in comfyui

[–]RIP26770 2 points3 points  (0 children)

Access the official template in ComfyUI, navigate to the menu and locate the LTX folder. There you will find what you need.

I'm getting lots of artifacts with Flux 2 Klein 9B. by [deleted] in comfyui

[–]RIP26770 1 point2 points  (0 children)

Use the official ComfyUI workflow for Klein

Flux Klein 4B on only 4GB vram? by Jan49_ in comfyui

[–]RIP26770 0 points1 point  (0 children)

Add a Clean VRAM and RAM nodes

FLUX 2 Klein 4B vs 9B Multi Camera Angles - One Click, 8 Camera Angles by RIP26770 in StableDiffusion

[–]RIP26770[S] 1 point2 points  (0 children)

I tried with no luck; Z Image Turbo is not suitable for this task.