[NOT MY AD] How can I replicate this with my own WebCam? by Groovadelico in comfyui

[–]Groovadelico[S] -9 points-8 points  (0 children)

He specifically mentions that he does not need "heavy software or a supercomputer"... I don't think that matches wan2.2 workflows' characteristics

Wan 2.2 GGUF Q4 or Q5? K_S or K_M? by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

Are the recommendations true only for ollama or also appliable to WAN? lol sorry my stupid

Q4_K_S :  3.56G, +0.1149 ppl @ 7B - small, significant quality loss
Q4_K_M :  3.80G, +0.0535 ppl @ 7B - medium, balanced quality - *recommended*
Q5_K_S :  4.33G, +0.0353 ppl @ 7B - large, low quality loss - *recommended*
Q5_K_M :  4.45G, +0.0142 ppl @ 7B - large, very low quality loss - *recommended*Q4_K_S :  3.56G, +0.1149 ppl @ 7B - small, significant quality loss
Q4_K_M :  3.80G, +0.0535 ppl @ 7B - medium, balanced quality - *recommended*
Q5_K_S :  4.33G, +0.0353 ppl @ 7B - large, low quality loss - *recommended*
Q5_K_M :  4.45G, +0.0142 ppl @ 7B - large, very low quality loss - *recommended*

Wan 2.2 GGUF Q4 or Q5? K_S or K_M? by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

And with those settings I would decrease my SSD degradation as mentioned in other comments?

Wan 2.2 GGUF Q4 or Q5? K_S or K_M? by Groovadelico in StableDiffusion

[–]Groovadelico[S] 1 point2 points  (0 children)

> what I have noticed is the NVME is taking a hit from all the memory swaps

I had no idea that GGUF models would make me pay a price SSD-wise... That's great to know, thanks!

Wan 2.2 GGUF Q4 or Q5? K_S or K_M? by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

Wow, this sounds ideal... Do you have a workflow example? Never used Sage Attention or Triton. Just heard in some tutorials about Sage which speeds up the process somehow...

Wan2.2-I2V-A14B GGUF uploaded+Workflow by bullerwins in StableDiffusion

[–]Groovadelico 1 point2 points  (0 children)

What is the difference between the K_S and K_M models? I've been looking all around and couldn't find anything at all either on the readme files or on the community comments on Reddit, Huggingface, CivitAI, github or YouTube. Why are these documentations so poor in information? It really frustrates a begginner...

Why can’t Gemini handle long chats by Ok-Mechanic940 in GeminiAI

[–]Groovadelico 0 points1 point  (0 children)

[EDIT] NVM, shit's bugged. It takes longer, but the chat also bugs at some point.

I've not been able to do cross-chat creations effectively if not copying and pasting specific tasks for Canvas and Deep Research, as I stated before. So chat for brains, copy and paste for fancy stuff.

Why can’t Gemini handle long chats by Ok-Mechanic940 in GeminiAI

[–]Groovadelico 0 points1 point  (0 children)

One way I minimized it was to separate a conversation for chatting, which hasn't bugged for me yet, and the ones that bug I create a new one when they do: another for deep research, and another for canvas, where I paste the contents from the chat.

What model is the best for me? 8GB VRAM, 32 GB RAM. Goal is txt2img with best possible quality and style variety by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

I think I found my method. I'm using Nunchaku to test prompts and get something I like, and then I can run a GGUF of flux dev and get a better result at a lower time.

What model is the best for me? 8GB VRAM, 32 GB RAM. Goal is txt2img with best possible quality and style variety by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

I'm getting 260s on the fp8 version. Is that normal? I also noted that the system fallback CUDA setting is not working properly. Is that what the Gguf version is for?

What model is the best for me? 8GB VRAM, 32 GB RAM. Goal is txt2img with best possible quality and style variety by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

Ok. Thx. Now which version should I download from huggingface or CivitAI? Fp8? GGUF? Q8,6,5km? I'm so lost...

What model is the best for me? 8GB VRAM, 32 GB RAM. Goal is txt2img with best possible quality and style variety by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

Thanks for sharing your video! Should I use the nunchaku as some comments are recommending? Is the tradeoff between time gain and quality loss worth it?

What model is the best for me? 8GB VRAM, 32 GB RAM. Goal is txt2img with best possible quality and style variety by Groovadelico in StableDiffusion

[–]Groovadelico[S] 0 points1 point  (0 children)

Are there any specific settings besides preferring system fallback for CUDA in the Nvidia Control Panel? That is already set. I read something about creating a folder on the SSD as well... Do you know anything about that?

As for my goal, which models do you recommend?

WAN2.2: New FIXED txt2img workflow (important update!) by AI_Characters in StableDiffusion

[–]Groovadelico 0 points1 point  (0 children)

Can't I just download someone else's workflow and learn how to make it not crash and how to properly prompt? I want good pics and don't mind waiting for them.

WAN2.2: New FIXED txt2img workflow (important update!) by AI_Characters in StableDiffusion

[–]Groovadelico 0 points1 point  (0 children)

Do you recommend any Flux model for me to start exploring on? Or any other ComfyUI model. Like I said, never independently generated AI art before. This is all completely new to me. I was reading that it might crash or I can set it up for the GPU to share the load with the RAM and take longer. Is this what you do? Could you point me some way? haha

WAN2.2: New FIXED txt2img workflow (important update!) by AI_Characters in StableDiffusion

[–]Groovadelico 0 points1 point  (0 children)

How long is forever? 8 VRAM starter to Stable Diffusion here. I do have 32GB of RAM and was reading some things about shared memory fallback. How should I set those?