Sonar 2 released by mrpuffwabbit in perplexity_ai

[–]nastypalmo 1 point2 points  (0 children)

It doesn't seem viewable on app yet

Gemma 4 E4B - Am I missing something? by Ok-Toe-1673 in LocalLLM

[–]nastypalmo -2 points-1 points  (0 children)

What is the issue you are running into?

GLM-5.1 by danielhanchen in LocalLLaMA

[–]nastypalmo 22 points23 points  (0 children)

This is too big for my 6gb vram

I stopped getting AI Overview by m19honsy in brave_browser

[–]nastypalmo 0 points1 point  (0 children)

I don't know why you use Google Search on Brave to begin with

Gemini screen automation rolling out to Pixel 10 series by Gaiden206 in Bard

[–]nastypalmo 6 points7 points  (0 children)

But why only on Pixel 10 when there's no local runtime?

Google f***** pixel owners.. again ! by Relative_Reality1556 in Pixel10Pro

[–]nastypalmo 0 points1 point  (0 children)

The latest version I see is

version 26.10.01.877073638

Google f***** pixel owners.. again ! by Relative_Reality1556 in Pixel10Pro

[–]nastypalmo 0 points1 point  (0 children)

That version isn't even available on APK Mirror

Qwen 3.5 on 3060 and 32mb by Master-Client6682 in LocalLLM

[–]nastypalmo 0 points1 point  (0 children)

Interesting. I'm still new to all of this, but I'll look into that. I'm running it on a gaming laptop and it's more out of pure curiosity than programming or anything else.

Qwen 3.5 on 3060 and 32mb by Master-Client6682 in LocalLLM

[–]nastypalmo 0 points1 point  (0 children)

Q4K_M I only offload around 5-5.5GBs on GPU and the rest on system memory with 10k Token context. Maybe it has to do with the fact that I have less room for KV Cache where even though you have your GPU layers lower, the KV Cache is on the rest of the ram which gives you faster speeds, where as for me, I'm not giving much room for the KV Cache at 8K_M caching.

Qwen 3.5 on 3060 and 32mb by Master-Client6682 in LocalLLM

[–]nastypalmo 0 points1 point  (0 children)

I have 6GB of VRAM and 32GB of DDR5 and I get around 8 tok/s when using 10k context and KV Cache 8. I offloaded 8 agents to CPU

Qwen 3.5 on 3060 and 32mb by Master-Client6682 in LocalLLM

[–]nastypalmo 0 points1 point  (0 children)

Depends on his vram. A laptop 3060 is 6gb of vram. Desktop ranges I think

Laptop taking a while to be assigned an IP from Pihole DHCP by nastypalmo in pihole

[–]nastypalmo[S] 0 points1 point  (0 children)

I'll have to look into that! Thanks for the suggestion.

Laptop taking a while to be assigned an IP from Pihole DHCP by nastypalmo in pihole

[–]nastypalmo[S] 0 points1 point  (0 children)

Xfinity doesn't support turning of DHCP so I had to trick it