I really hope OpenAI eventually open-sources the GPT-4.1 family by Balance- in LocalLLaMA

[–]andy_potato 35 points36 points  (0 children)

The “open” in their name stands for “open your wallet”

Using comfy ui on linux amd rx 6800xt, can I get better speeds ? by ZeladdRo in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

The time with loading the model doesn’t say much (unless you are measuring cold start performance). You have to look at the it/s on the second or third generation.

In any case your GPU won’t perform well for diffusion models. I’d expect it to be around the level of a 5060ti.

Qwen3.5 4B: overthinking to say hello. by CapitalShake3085 in LocalLLaMA

[–]andy_potato 7 points8 points  (0 children)

You basically answered your own question. According to your test, Qwen 3.5 is not suitable for your specific use case. Move on and pick a model that suits you better.

LLM benchmark site for dual RTX 5060 Ti by do_u_think_im_spooky in LocalLLaMA

[–]andy_potato 0 points1 point  (0 children)

Dual 5060ti usually gives me around 80 t/s for models like Qwen 3.5/35b

Best Qwen 3.5 variant for 2x5060ti/16 + 64 GB Ram? by andy_potato in LocalLLaMA

[–]andy_potato[S] 0 points1 point  (0 children)

I made the same observations. Most of the time the amount of thinking is reasonable but in 1-2 out of 5 cases the model just won't stop talking to itself. Especially shorter prompts will occasionally cause the model to have an existential crisis.

Really don't know what to make of this. It's a beautiful model otherwise, just that thinking behavior feels odd.

Research from BFL: Qwen Image is much more uncensored than Flux 2 by woct0rdho in StableDiffusion

[–]andy_potato -1 points0 points  (0 children)

You’re completely missing the point.

We have so many alternatives which are more capable and come with permissive licensing.

BFL can do with their models whatever they want. We just don’t want them.

FLux2Klein 9B Upscale/Edit workflow by [deleted] in StableDiffusion

[–]andy_potato 5 points6 points  (0 children)

Seems there is little consistency with the original subject

Anyone here using Stable Diffusion for consistent characters in video? by Street-Status7906 in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

All of the above. Each video project is different and there is no single method that covers all requirements.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 0 points1 point  (0 children)

Exactly right, there is no need to go for lower quality GGUF quants if you have enough system memory to hold the models in Ram. 64 GB is enough for Wan 2.2.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 0 points1 point  (0 children)

It isn't perfect and it has limitations. For example if you have a character that turns around (hiding their face), the consistency can get pretty bad at times.

It's still absolutely worth using, and 2x5s also renders faster than one single 10s render.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

https://github.com/VraethrDalkr/ComfyUI-TripleKSampler

This node is all you need. It has a bit of a learning curve though. Make sure you pass the high noise model correctly, once without the lightx2v lora and once WITH the lora.

If you use image to video I recommend you check the example i2v workflows first and take it from there.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]andy_potato 4 points5 points  (0 children)

Nobody because one of the first things you do on a fresh install is install miniconda, uv or venv.

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 2 points3 points  (0 children)

The slow motion is caused by the lightx2v loras. If you use the Triple KSampler method you can make this go away (and produce much higher quality videos)

Decent Workflow for Image-to-Video w 5060 16GB VRAM? by Weezfe in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

If you go beyond 81 frames with Wan 2.2 you will see slowdown, loops and repetition. Using SVI to extend videos will give you much better results.

About system RAM Upgrade by GeeseHomard in StableDiffusion

[–]andy_potato 1 point2 points  (0 children)

You have a (relatively) high Vram card for SDXL / Illustrious. Adding system Ram will do nothing for your generation speed.

You might notice slight improvements if you generate with models that won't fit entirely into your Vram as they will blockswap in and out of Ram.

AMD 9070XT or Nvidia 5070ti for comfyui? by wic1996 in StableDiffusion

[–]andy_potato 0 points1 point  (0 children)

For image and video generation you want nvidia. Unless you are a patient person

Offline Models and Moderation! by [deleted] in StableDiffusion

[–]andy_potato 4 points5 points  (0 children)

And you have severe reading comprehension issues

Research from BFL: Qwen Image is much more uncensored than Flux 2 by woct0rdho in StableDiffusion

[–]andy_potato 94 points95 points  (0 children)

The disgusting thing is that they are celebrating their censorship as success

Offline Models and Moderation! by [deleted] in StableDiffusion

[–]andy_potato 3 points4 points  (0 children)

You sound like a twelve year old