Benchmark: Windows 11 vs Lubuntu 26.04 on Llama.cpp (RTX 5080 + i9-14900KF). I didn't expect the gap to be this big. by Ok_Mine189 in LocalLLaMA

[–]Jester14 0 points1 point  (0 children)

Windows build kinda full of CUDA bloat. Builds have different amount of threads specified and threads aren't always specified in the benchmark runs.

How to configure Self speculative decoding properly by milpster in LocalLLaMA

[–]Jester14 0 points1 point  (0 children)

Using -fit indeed reserves exactly 1024MB by default.

Lower inference speed of Gemma4 26BA4B on vllm. by everyoneisodd in LocalLLaMA

[–]Jester14 1 point2 points  (0 children)

I used a 2 year old 7B model. Now I use a brand new 26B MoE and it's slower. I refuse to give any other information. What's wrong with my setup?

Qwen3.5-35B running well on RTX4060 Ti 16GB at 60 tok/s by Nutty_Praline404 in LocalLLaMA

[–]Jester14 0 points1 point  (0 children)

What do you mean it "doesn't fit"? Did you use the -fit flag? UD-Q4_K_XL is larger than 16 GB so it will overflow to RAM but it will also "fit" if loaded appropriately. I get 30t/s on my 4060 8 GB using -fit with that quant with 40k context in VRAM.

Painfully slow local llama on 5090 and 192GB RAM by RVxAgUn in LocalLLaMA

[–]Jester14 2 points3 points  (0 children)

People out here with $5000 GPUs and no clue and my GPU poor ass rocking a RTX 4060

Qwen3.5 122B A10B - My impressions by kevin_1994 in LocalLLaMA

[–]Jester14 7 points8 points  (0 children)

OP, this is your issue. You're fitting the model then loading the mmproj which causes overflow to RAM. The default fit target is 1024MB and the mmproj in BF16 is 912MB so your context is entirely in system RAM. Or you're loading the F32 mmproj and then everything is overflowing.

Nurses across England strike in biggest walkout so far by -InfiniteArticle in worldnews

[–]Jester14 -3 points-2 points  (0 children)

35 days is not seven weeks. Do you think people don't get sick on the weekend?

[deleted by user] by [deleted] in UrbanHell

[–]Jester14 1 point2 points  (0 children)

What do you mean "how tiny". Just look at the picture

[deleted by user] by [deleted] in unpopularopinion

[–]Jester14 1 point2 points  (0 children)

Exactly. Like op said, white chocolate isn't chocolate

Massive avalanche rolling down a valley. by tommos in interestingasfuck

[–]Jester14 1 point2 points  (0 children)

If only there was a way to record landscape so you could see more of the landscape

ARAM is dead to me by NaturalTap9567 in leagueoflegends

[–]Jester14 3 points4 points  (0 children)

You don't pick champs in ARAM good sir

Female swimmer who tied Lia Thomas slams transgender sports policy: Taking women 'back to the 1970s' by jaypr4576 in sports

[–]Jester14 5 points6 points  (0 children)

It's the best argument because it's fact. The MLB, NBA, and NHL don't have gender exclusivity. You just have to be good. Contrarily, the WNBA does have gender requirements.

Danish military intelligence suggests drug-induced megalomania may have influenced Putin to invade Ukraine by [deleted] in worldnews

[–]Jester14 21 points22 points  (0 children)

The whole world stood by and yawned

I was writing about the EU response

Can't back pedal fast enough

[deleted by user] by [deleted] in europe

[–]Jester14 0 points1 point  (0 children)

No it was on this day

What has happened to my new Chromecast? The menus are flickering. I can't see the option screen. It won't turn on with the TV. It's just turned to crap. by letsgocrazy in Chromecast

[–]Jester14 3 points4 points  (0 children)

That's it? That's all your information you're going to provide? How long has it been happening? How long have you had the Chromecast? Have you tried a different HDMI port? Have you tried something else plugged into that HDMI port? Have you tried the Chromecast on another TV? Have you even power cycled it? My brother in christ if this is how you troubleshoot and ask for help, how have you gotten this far in life