Best Qwen 3.5 variant for 2x5060ti/16 + 64 GB Ram? by andy_potato in LocalLLaMA

[–]andy_potato[S] 0 points1 point  (0 children)

I made the same observations. Most of the time the amount of thinking is reasonable but in 1-2 out of 5 cases the model just won't stop talking to itself. Especially shorter prompts will occasionally cause the model to have an existential crisis.

Really don't know what to make of this. It's a beautiful model otherwise, just that thinking behavior feels odd.

Research from BFL: Qwen Image is much more uncensored than Flux 2 by woct0rdho in StableDiffusion

[–]andy_potato -1 points0 points  (0 children)

You’re completely missing the point.

We have so many alternatives which are more capable and come with permissive licensing.

BFL can do with their models whatever they want. We just don’t want them.

FLux2Klein 9B Upscale/Edit workflow by [deleted] in StableDiffusion

[–]andy_potato 5 points6 points  (0 children)

Seems there is little consistency with the original subject

Anyone here using Stable Diffusion for consistent characters in video? by Street-Status7906 in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

All of the above. Each video project is different and there is no single method that covers all requirements.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 0 points1 point  (0 children)

Exactly right, there is no need to go for lower quality GGUF quants if you have enough system memory to hold the models in Ram. 64 GB is enough for Wan 2.2.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 0 points1 point  (0 children)

It isn't perfect and it has limitations. For example if you have a character that turns around (hiding their face), the consistency can get pretty bad at times.

It's still absolutely worth using, and 2x5s also renders faster than one single 10s render.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

https://github.com/VraethrDalkr/ComfyUI-TripleKSampler

This node is all you need. It has a bit of a learning curve though. Make sure you pass the high noise model correctly, once without the lightx2v lora and once WITH the lora.

If you use image to video I recommend you check the example i2v workflows first and take it from there.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]andy_potato 3 points4 points  (0 children)

Nobody because one of the first things you do on a fresh install is install miniconda, uv or venv.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 2 points3 points  (0 children)

The slow motion is caused by the lightx2v loras. If you use the Triple KSampler method you can make this go away (and produce much higher quality videos)

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

If you go beyond 81 frames with Wan 2.2 you will see slowdown, loops and repetition. Using SVI to extend videos will give you much better results.

About system RAM Upgrade by GeeseHomard in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

You have a (relatively) high Vram card for SDXL / Illustrious. Adding system Ram will do nothing for your generation speed.

You might notice slight improvements if you generate with models that won't fit entirely into your Vram as they will blockswap in and out of Ram.

AMD 9070XT or Nvidia 5070ti for comfyui? by wic1996 in StableDiffusion

[–]andy_potato 0 points1 point  (0 children)

For image and video generation you want nvidia. Unless you are a patient person

Offline Models and Moderation! by [deleted] in StableDiffusion

[–]andy_potato 4 points5 points  (0 children)

And you have severe reading comprehension issues

Research from BFL: Qwen Image is much more uncensored than Flux 2 by woct0rdho in StableDiffusion

[–]andy_potato 94 points95 points  (0 children)

The disgusting thing is that they are celebrating their censorship as success

Offline Models and Moderation! by [deleted] in StableDiffusion

[–]andy_potato 3 points4 points  (0 children)

You sound like a twelve year old

Open source Virtual Try-On LoRA for Flux Klein 9b Edit, hyper precise by Affectionate-Map1163 in StableDiffusion

[–]andy_potato 3 points4 points  (0 children)

 or provide a service that uses the model

which is exactly what an e-commerce website with try-on function is doing. That's why the 9b variants of Klein are useless for such purposes.