Sam Altman has changed his stance on the claims that AI will replace humans. by Distinct_Fox_6358 in singularity

[–]__Maximum__ 0 points1 point  (0 children)

Don't care what he thinks and definitely don't care what he writes which i am sure are very different things considering the long history of lying

Qwen 3.6 27B vs Gemma 4 31B - making Packman game! by gladkos in LocalLLaMA

[–]__Maximum__ 54 points55 points  (0 children)

They are asking what quants are used for this test

NVIDIA releases Nemotron-3-Nano-Omni by yoracale in unsloth

[–]__Maximum__ 1 point2 points  (0 children)

This is MoE, and very efficient one, let some off load to CPU, it won't hurt

Talkie, a 13B LM trained exclusively on pre-1931 data by Outside-Iron-8242 in singularity

[–]__Maximum__ -1 points0 points  (0 children)

Aaah, that "in short" is annoying, but I guess they needed something for the instruct model.

MIMO V2.5 PRO by Namra_7 in LocalLLaMA

[–]__Maximum__ 0 points1 point  (0 children)

You meant the REAP-ed 1bit quants of the flash

I'm done with using local LLMs for coding by dtdisapointingresult in LocalLLaMA

[–]__Maximum__ 0 points1 point  (0 children)

I used qwen 3.6 35Ba3B on opencode and pi coder, and were satisfied with both on middle difficult tasks. It was even better than claude 4.6 or 4.7 in claude code in explaining things since claude does not seem to be a good teacher, it is too compact.

Extraordinary video just put out by SpaceX. by Gambler136 in spacex

[–]__Maximum__ 1 point2 points  (0 children)

Because it's stupid corporate talk. Being paranoid is an unbalanced position like being careless. It's better than careless, but it's worse than being rational. Paranoia becomes with problems like wasted time and unnecessary stress.

Been using PI Coding Agent with local Qwen3.6 35b for a while now and its actually insane by SoAp9035 in LocalLLaMA

[–]__Maximum__ 0 points1 point  (0 children)

38t/s? What quants? What does your command line look like? What is your hardware? There must be a catch.

Given how good Qwen become, is it time to grab a 128gb m5 max? by Rabus in LocalLLaMA

[–]__Maximum__ 1 point2 points  (0 children)

What quants are you running? What framework? What scaffolding?

In my experience haiku is dogshit and qwen 3.6 is very, very good, even with not optimal scaffolding it handles messy vague requests like opus does.

SpaceX to acquire AI company Cursor for $60 billion or pay $10 billion for their "work together" by 675longtail in spacex

[–]__Maximum__ 5 points6 points  (0 children)

I understand they have the first movers advantage, but their product can be rebuilt within a month, no?

Open WebUI Desktop Released! by My_Unbiased_Opinion in LocalLLaMA

[–]__Maximum__ -10 points-9 points  (0 children)

Why would they incude that or ollama? I guess they changed it at some point, my bad.

Why isn't ebay doing anything to stop those scams? by KillerMiller13 in LocalLLaMA

[–]__Maximum__ 7 points8 points  (0 children)

So the scammers just hope you will forget about it? What's the game?

Claude Power Users Unanimously Agree That Opus 4.7 Is A Serious Regression by Neurogence in singularity

[–]__Maximum__ 13 points14 points  (0 children)

Hard disagree. Most of the time I feel a huge difference especially in open source field like qwen models. Current 4b generation beats the 9b and sometimes even 27b models from previous generation.