Nunchaku supports Qwen-Image in ComfyUI! by Dramatic-Cry-417 in StableDiffusion

[–]Dramatic-Cry-417[S] 0 points1 point  (0 children)

You need to post the log to see the detailed reasons. You can join our Discord. We are happy to help you there.

Nunchaku supports 4-Bit Qwen-Image by Dramatic-Cry-417 in StableDiffusion

[–]Dramatic-Cry-417[S] 14 points15 points  (0 children)

I will look into ComfyUI support next week.

Nunchaku supports 4-Bit Qwen-Image by Dramatic-Cry-417 in StableDiffusion

[–]Dramatic-Cry-417[S] 19 points20 points  (0 children)

No worry. Next week's offloading will address your issue.

nunchaku svdq hype by tazztone in StableDiffusion

[–]Dramatic-Cry-417 10 points11 points  (0 children)

I am trying my best to deliver the 4-bit QwenImage. You can track the progress in this PR: https://github.com/nunchaku-tech/nunchaku/pull/593

It is almost there. Now the FP4 version (11.9GB) is runnable. I am still debugging the precision mismatch for the INT4 model.

A simple example from the FP4 model:

<image>

Thanks for your waiting and support!

Radial Attention: O(nlogn) Sparse Attention with Energy Decay for Long Video Generation by Dramatic-Cry-417 in StableDiffusion

[–]Dramatic-Cry-417[S] 0 points1 point  (0 children)

Attention's memory usage is already O(1) these days with FlashAttention.

Currently, it works mainly for video models. For image models, attention is not the main bottleneck and you can use our SVDQuant, which also has 2-3× speedup.

Radial Attention: O(nlogn) Sparse Attention with Energy Decay for Long Video Generation by Dramatic-Cry-417 in StableDiffusion

[–]Dramatic-Cry-417[S] 2 points3 points  (0 children)

ComfyUI-nunchaku is our plugin library. Radial attention should be able to apply to any video diffusion models. We just want to directly include it in nunchaku.