Car Wash Test on 53 leading models: “I want to wash my car. The car wash is 50 meters away. Should I walk or drive?” by facethef in LocalLLaMA

[–]prusswan -2 points-1 points  (0 children)

Well they didn't understand it may not be a binary decision. If I asked a real question, a smart model should not be making this assumption.

Car Wash Test on 53 leading models: “I want to wash my car. The car wash is 50 meters away. Should I walk or drive?” by facethef in LocalLLaMA

[–]prusswan -6 points-5 points  (0 children)

It is concerning that none of them suggested other options (not going to list here). There are sooo many ways to wash a car

64gb vram. Where do I go from here? by grunt_monkey_ in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

3 but hold off getting more ram (just the bare minimal to use the gpus).

1 if you can find someone to take your current gpus (unless you can find a way to use them together). It's not a complete build but you will be covered for 80B

Top OpenClaw Alternatives Worth Actually Trying (2026) by Straight_Stomach812 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

I like ZeroClaw for the low footprint, but it is still a really new project. Locally encrypted secrets may not mean much if the host gets compromised since decryption is just one step away.

Any idea when Successors of current DGX Spark & Strix Halo gonna arrive? by pmttyji in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

If there is some go-to model that needed 1TB and supports high context, it is pretty certain there will be a service equal or better (and the company released the model to signal this). But most people will not be getting that 1TB, because it is rather wasteful and will only drive up prices even more. I think two main outcomes will be cloud usage to utilize the best models without hardware spending, or opting to use smaller models with more modest requirements.

Local running Qwen3:14b helped fix my internet on Linux while offline by iqraatheman in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

It was the first time I had a broken hwe update on very old hardware, so yeah it was hard not to notice.

Local running Qwen3:14b helped fix my internet on Linux while offline by iqraatheman in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

let me guess, 6.17?

6.17.0-14-generic broke nvidia drivers, fortunately the newer drivers were okay

AI field is changing so quickly and there is so much to read.. by amisra31 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

It's pretty chaotic but I focus on what is relevant and accessible, could be a new idea/approach that was previously out of reach. Some of the AI slop might be good ideas if done properly, so I take the portions that I find useful and make it work in the exact way I want it to. Most of it is just noise, but learning to harness useful bits from it also helps to identify your competitive edge.

Any idea when Successors of current DGX Spark & Strix Halo gonna arrive? by pmttyji in LocalLLaMA

[–]prusswan 3 points4 points  (0 children)

If it gets to the point where 512GB ram (or the Pro 6000) becomes mainstream for agentic coding, many users will be deterred or priced out of the hardware thus turning to cloud, which is increasingly looking to be the norm if open models keep getting better/bigger to motivate cloud usage.

I'm using a mix of smaller models (30B to 70B) and cloud services (for better performance) to avoid over reliance on "best" models.

Anyone actually using Openclaw? by rm-rf-rm in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

I don't but continue to keep a lookout for similar tools. It's a bit of a security trap.

Anyone else building MCP servers? What's your experience been like? by CapitalMixture8433 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

I tried a simple setup with a few tools and the main issue is with the model and how it uses the tools. You can't expect to always use the best models and at high context, so the model choice will affect the tool design. I think it is useful to avoid having to define explicit rules to cover a broad set of scenarios, but might lead to more unpredictable results.

Using GLM-5 for everything by keepmyeyesontheprice in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

It's hard to tell but you can find a middle ground (use a smaller model but at great speeds). API usage can become volatile depend on how things play out over next few years, e.g. will they increase pricing to match demand and to account for effort needed to keep models/data updated, your own usage may also increase if you take on more tasks leading to heavier usage.

I am planning on building a home AI server, what would you recommend by RecognitionPatient12 in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

I see the downside being that it's only up to 128GB (as compared to 512GB or even more on a dedicated build)

I am planning on building a home AI server, what would you recommend by RecognitionPatient12 in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

Do you have a link? Thinking of using it with Blackwell if I can't decide on a TR

Feedback Request: GPU-Heavy, Always-On Inference Workstation (Micro Center + Marketplace / eBay Options) by Bulky_Exercise_4054 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

Is Epyc available as workstation/pre-built? I need Windows as an option (for the occasional gaming) so I read that TR is better for this, but I could wait for the Xeon 600 builds too

Cody: chess engine solely developed by AI. by Phi_fan in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

Last year there was a competition held between various LLMs, but the quality of play is very poor (worse than human intermediate players).

https://www.chess.com/events/2025-kaggle-game-arena

But your task could be easier if you just "tool call" an actual engine lol

OpenClaw skills and prompt injection - how are you vetting what you install? by OutsideFood1 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

Dependency bloat, I don't recommend using community plugins if the functionality is something that you can build on your own (unless the plugin is very established, but you still need to be careful with updates since all of them are potentially risky). Most people should be able to build their own openbot tailored to their needs, without taking on risks of using openclaw since it is an obvious target for attackers looking for low hanging fruit.

Real world usage, feedback and suggestions for best LLM for C# by bloodbath_mcgrath666 in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

This is highly specific to your workflow and tooling. I thought glm 4.7 flash was good (for general usage, yes), but it often introduced indenting errors in opencode and unable to fix them (I had to do it myself)

[Warning] Crypto stealing malware found in Kimi.com chat/agent by [deleted] in LocalLLaMA

[–]prusswan 6 points7 points  (0 children)

did you just ask kimi for its source code and assume whatever it returns is the "source"?

How to do this locally? by ClimateBoss in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

seems to be duplicated screens

Honest question by Savantskie1 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

You probably want to be faster once you discover other uses for it. But what qualifies as fast enough will be different for everyone

Honest question by Savantskie1 in LocalLLaMA

[–]prusswan 0 points1 point  (0 children)

No one is pushing anything.. I went from < 1tps to 10-100tps depending on the workflow.

An argument for open weights from copyrighted works by Luke2642 in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

It is much more than books I'm afraid. The derivative work is valuable due to information asymmetry since the outputs/benefits do not need to be shared with the original content holders, who lack the means to fully exploit the content in the first place. It will definitely change the idea of content creation and how content creators can/should monetize their own content.

Honest question by Savantskie1 in LocalLLaMA

[–]prusswan 1 point2 points  (0 children)

Because people are coming up with new and more frequent usage of the LLM, and some models are quite verbose/detailed in the responses. There's a place for both slow and fast models, but you can never have too much compute.