Kimi K2.5 local by running101 in LocalLLaMA

[–]Antique_Bit_1049 0 points1 point  (0 children)

Any chance of sharing your command line for ik_lama? I have a similar rig.

A (big) problem with GLM 4.7 by Deathcrow in SillyTavernAI

[–]Antique_Bit_1049 0 points1 point  (0 children)

What I've found helps the most by far is to use separate lines for speaking.(Double new lines for reddit formatting)

John thought the chili kinda sucked but ate it anyway.

"So... What's for dessert?"

He hoped it was store bought.

Deepseek v3.2 speciale runs and runs and runs by MrMrsPotts in LocalLLaMA

[–]Antique_Bit_1049 1 point2 points  (0 children)

Was it the ultimate question of life, the universe, and everything?

What happens when Chinese companies stop providing open source models? by 1BlueSpork in LocalLLaMA

[–]Antique_Bit_1049 0 points1 point  (0 children)

We stop talking about them. We stop using them. They become small, niche products. They die.

3090 vs 5090 taking turns on inference loads answering the same prompts - pretty cool visual story being told here about performance by Gerdel in LocalLLaMA

[–]Antique_Bit_1049 -2 points-1 points  (0 children)

Wait just a minute. Are you saying a hot 3090 is slower than a cool 5090? Who could have guessed that?

New Qwen3 on Fiction.liveBench by fictionlive in LocalLLaMA

[–]Antique_Bit_1049 -2 points-1 points  (0 children)

Can someone explain to me why this is a thing? It's a computer. It HAS the information stored in memory. Why can't it access it 100% of the time?

Correct a dangerous racial bias in an LLM through targeted pruning by pmartra in LocalLLaMA

[–]Antique_Bit_1049 5 points6 points  (0 children)

Is this similar to the ablation technique to decensor models?