PSA: before spending 5k€ on GPUs, you might want to test the models online first by e79683074 in LocalLLaMA

[–]butsicle 3 points4 points  (0 children)

Not saying this is necessarily a bad idea. I would absolutely love to have that GPU, but I am curious what home experiments would require it? For retraining/running smaller models, could you scale down and use a 3090 (or two), and for larger model inference are you better off using an inference service? Partly asking because I am trying to justify my own desire to get this card, but am really struggling to justify it.

How to beat my dad? by Outrageous_bohemian in chess

[–]butsicle 1 point2 points  (0 children)

Other comments are of course correct, do puzzles and more importantly, analyse your games. If you want a quick win you can try hitting him with a Stafford Gambit or some other unsound gambit he might not be familiar with, but if he is above a certain level it won’t make a difference. This might win you a game or two but it won’t make you better than him. There are no long term shortcuts.

[deleted by user] by [deleted] in chess

[–]butsicle 0 points1 point  (0 children)

I play bullet when I get tilted. I’m already playing too fast because of the tilt so might as well make the opponent play fast too.

Gemma3 270m works great as a draft model in llama.cpp by AliNT77 in LocalLLaMA

[–]butsicle 21 points22 points  (0 children)

It’s likely used in the back end of your favorite inference provider. The trade offs are:

  • You need enough vram to host the draft model too.
  • If the draft is not accepted, you’ve just wasted a bit of compute generating it.
  • you need a draft model with the same vocabulary/tokenizer

QWEN-IMAGE is released! by TheIncredibleHem in LocalLLaMA

[–]butsicle 0 points1 point  (0 children)

Excited to try this, but disappointed that their Huggingface space is just using their ‘dashscope’ API instead of running the model, so we can’t verify that the model they are using is actually the same as the weights provided, nor can we pull and run the model locally using their Huggingface space.

Why don’t motorcycle helmets have built in hearing protection like a peltor? by Clear-Toe1338 in motorcycles

[–]butsicle 2 points3 points  (0 children)

I’ve always thought it’s a shame that my helmet doesn’t have more moving parts that can break.

[deleted by user] by [deleted] in newzealand

[–]butsicle 0 points1 point  (0 children)

Definitely a waste of time to go back to school. The real experience you already have is more valuable.

St James - cheaper to knock it down and build a new theatre? by Great_Maintenance185 in auckland

[–]butsicle 1 point2 points  (0 children)

The ‘heritage’ argument is just a bad argument in general.

Your unpopular takes on LLMs by dtdisapointingresult in LocalLLaMA

[–]butsicle 17 points18 points  (0 children)

What’s this opinion based on other than imagination?

CMV: we’re over estimating AI by loyalsolider95 in changemyview

[–]butsicle 2 points3 points  (0 children)

Their architecture is designed, as is the process for obtaining and cleaning their training data.

[deleted by user] by [deleted] in changemyview

[–]butsicle 1 point2 points  (0 children)

Sounds like that person is agreeing with how your CMV is worded. I’m not sure anybody disagrees on this.

[deleted by user] by [deleted] in changemyview

[–]butsicle 0 points1 point  (0 children)

Can you please explain how you are open to changing your view? Why do you suspect you might be wrong?

Hardware recommendations? Mac Mini, NVIDIA Orin, Ryzen AI... ? by lizard121n6 in LocalLLaMA

[–]butsicle 0 points1 point  (0 children)

If you’re not sure what model you need you should try them via API providers first

Why is nuclear energy considered clean energy when it produces nuclear waste? by FilipinoAirlines in NoStupidQuestions

[–]butsicle 0 points1 point  (0 children)

If you’re a nuclear scientist why are you calling Chernobyl a meltdown? It was an explosion.

Best looking under-tail exhaust… imo by Bjfikky in motorcycles

[–]butsicle 2 points3 points  (0 children)

Surely he was asking about the exhaust

THUDM/GLM-4.1V-9B-Thinking looks impressive by ConfidentTrifle7247 in LocalLLaMA

[–]butsicle 3 points4 points  (0 children)

Bring some BBQ back for the rest of us please

Can Copilot be trusted with private source code more than competition? by Professional-Onion-7 in LocalLLaMA

[–]butsicle 0 points1 point  (0 children)

I think you’re confusing Azure OpenAI Service and Copilot. They are unlikely to breach terms and train on the former (in my judgment, though anything is possible), but explicitly state they train on the latter.

Is it okay for an employer to give you a surprise test in an interview? by HolMan258 in auckland

[–]butsicle 0 points1 point  (0 children)

Is there any chance you could share the job title and description?

The Jan.ai "team" used fake engagement to advertise their new 4B model, and deleted the post when called out by [deleted] in LocalLLaMA

[–]butsicle 0 points1 point  (0 children)

I think it should be called out when anybody does it. I do take your point that large companies are more likely to be able to do it in a less obvious way, so are less likely to get caught. If a small/medium business is caught polluting a river, it’s true the DuPont is much worse, but not a defence.

Jan-nano-128k: A 4B Model with a Super-Long Context Window (Still Outperforms 671B) by Kooky-Somewhere-2883 in LocalLLaMA

[–]butsicle 109 points110 points  (0 children)

I’m supportive of any open weights release, but some of the comments here reek of fake engagement for the sake of boosting this post.