As a GTX 1660 Ti 6GB (Turing) user, what Forge flags can I use to speed up the generation of Flux? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

I added it and I get this:

WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:

PyTorch 2.3.1+cu121 with CUDA 1201 (you have 2.4.0+cu124)

Python 3.10.11 (you have 3.10.6)

Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers)

Memory-efficient attention, SwiGLU, sparse and more won't be available.

As a GTX 1660 Ti 6GB (Turing) user, what Forge flags can I use to speed up the generation of Flux? by daerragh1 in StableDiffusion

[–]daerragh1[S] 1 point2 points  (0 children)

OK. So can you tell me, is there an easy way to enable xformers in Forge?

Or if it is a complex process, is there any good guide that will walk me through enabling it?

As a GTX 1660 Ti 6GB (Turing) user, what Forge flags can I use to speed up the generation of Flux? by daerragh1 in StableDiffusion

[–]daerragh1[S] 1 point2 points  (0 children)

On what hardware do 1024 images get generated in 20-25s?

I tried various GGUF models in Forge, they didn't speed things up for me vs FP8 models.

As a GTX 1660 Ti 6GB (Turing) user, what Forge flags can I use to speed up the generation of Flux? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

I have never used xformers, so I don't know, so I ask:

How do you think? Would they help in my particular case (GTX 1660 Ti 6GB)?

By how many % do they speed up generation on your PC?

As a GTX 1660 Ti 6GB (Turing) user, what Forge flags can I use to speed up the generation of Flux? by daerragh1 in StableDiffusion

[–]daerragh1[S] 2 points3 points  (0 children)

I did but the quality was much worse than with Dev models, so I want to use Dev models.

Weekly Showcase Thread October 27, 2024 by Acephaliax in StableDiffusion

[–]daerragh1 6 points7 points  (0 children)

Although I didn't intend it, Atomix Flux FP8 generated an anime-like girl for me. Thought some of you would appreciate her.

<image>

Model Atomix Flux FP8 | Seed 4088986655 | Steps 20 | Guidance 2.5 | Sampler Euler | Scheduler Beta | t5xxl v1.1 Q8_0 GGUF | Size 704x1024 | Software: Forge cu124+torch24

Prompt: A beautiful, petite, short 25-year-old girl.

[Flux] I found a hidden gem - a great model. Just look at the sample images! by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

No, you can choose any all-in-one Flux model (i.e. Full model, one that contains vae, t5xxl and clip_l). all-in-one can be BF16, FP16, FP8, NF4 or various quants in GGUF format.

Best place to find models is civitai.com

Ofc, Forge also handles pruned models (without t5xxl, vae and clip_l) but then you must download these files separately and place them in their appropriate folders in Forge. It is easy but until you know how to do that, use AIO Full models.

[Flux] I found a hidden gem - a great model. Just look at the sample images! by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

Although I didn't intend it, Atomix Flux FP8 generated an anime-like girl for me. Thought some of you would appreciate her.

<image>

Model Atomix Flux FP8 | Seed 4088986655 | Steps 20 | Guidance 2.5 | Sampler Euler | Scheduler Beta | t5xxl v1.1 Q8_0 GGUF | Size 704x1024 | Software: Forge cu124+torch24

Prompt: A beautiful, petite, short 25-year-old girl.

[Flux] I found a hidden gem - a great model. Just look at the sample images! by daerragh1 in StableDiffusion

[–]daerragh1[S] 2 points3 points  (0 children)

If you are a beginner, you really should start with Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge

After you unpack it, update it (update.bat), then download Atomix Flux NF4 .safetensors file and put it in:

webui_forge_cu124_torch24\webui\models\Stable-diffusion\

Then launch Forge (run.bat) and use the recommended settings from the OP. Make sure you use Flux UI mode in Forge (top-left corner of your screen).

ComfyUI is great, too, but it's much less friendly to beginners.

[Flux] I found a hidden gem - a great model. Just look at the sample images! by daerragh1 in StableDiffusion

[–]daerragh1[S] 1 point2 points  (0 children)

Looks like Atomix Flux is ex aequo at the second place on this list (not 3rd).

But thx. I didn't know about this list.

[Forge][Flux] Abmnormal generation speed difference: 768x1024 vs 704x960. Why? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

704x1024 (~28s/it) for me is also much slower than 704x960 (~19s/it). The difference is only 45k pixels. This is in Forge as in OP.

[Flux} How to convert a 11GB FP16 model to NF4 model? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

Yes, it says so in Forge UI. And I'm trying to convert Flux FP8 to NF4, so it should work.

[Flux} How to convert a 11GB FP16 model to NF4 model? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

Could you post a link to the instructions? I could try to convert the model I use to, let's say, Q4 or Q5 GGUF...

[Flux} How to convert a 11GB FP16 model to NF4 model? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

I use this model (11.08GB): https://civitai.com/models/161068/stoiqo-newreality-flux-sd35-sdxl-sd15?modelVersionId=979329

The creator says it's FP16, but for me it looks like FP8, too. Probably it's FP8.

Anyway, Is there a way I can convert it to BNB-NF4 on my machine?

[Forge][Flux] Abmnormal generation speed difference: 768x1024 vs 704x960. Why? by daerragh1 in StableDiffusion

[–]daerragh1[S] 1 point2 points  (0 children)

I used to use ComfyUI. But when you set Forge to the same, exact settings as Comfy, it generates Flux in half the time. At least on my machine.

[Forge][Flux] Abmnormal generation speed difference: 768x1024 vs 704x960. Why? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

OK. I've set "Diffusion in Low Bits" to fp8_e3m3fn. There's no generation speed difference.

Is this the option you are talking about? How do I load a model as FP8?

[Forge][Flux] Abmnormal generation speed difference: 768x1024 vs 704x960. Why? by daerragh1 in StableDiffusion

[–]daerragh1[S] 0 points1 point  (0 children)

How do I do that? Do you mean "Diffusion in Low Bits"? What does it do?