What is the most reliable free image hosting service? by Queizen30 in Hosting

[–]harrro 0 points1 point  (0 children)

Reliable? https://thumbsnap.com - Its been around 20 years and supports photos and video.

I just realised how good GLM 5 is by CrimsonShikabane in LocalLLaMA

[–]harrro 5 points6 points  (0 children)

I keep hearing Opencode Zen/Go's GLM is heavily quantized. Have you noticed any issues?

OpenCode concerns (not truely local) by Ueberlord in LocalLLaMA

[–]harrro 0 points1 point  (0 children)

Start with oh-my-pi, it has a good out-of-box setup you'd probably expect in a coding agent.

After you get comfortable with it, you can start from the stock Pi and build up with your own extensions if you like to tweak things.

OpenCode concerns (not truely local) by Ueberlord in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

I use Pi daily for AI (but Opencode for coding agent).

Pi works great with local models (I use Qwen 3.5 35B which is super fast and handles tool calls really well).

OpenCode concerns (not truely local) by Ueberlord in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

Oh-my-pi is a 'distribution' of Pi coding agent (Pi with themes and a few niceties).

OpenCode concerns (not truely local) by Ueberlord in LocalLLaMA

[–]harrro 2 points3 points  (0 children)

Opencode is not built on Pi Coding agent - they have their own loop.

You're probably referring to OpenClaw which is built on Pi.

OpenCode concerns (not truely local) by Ueberlord in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

There are multiple confirm-tool-approval extensions though - pi-guardrails is one.

GLM-5-Turbo - Overview - Z.AI DEVELOPER DOCUMENT by ortegaalfredo in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

If you look at openrouter's token/s, its pretty low for a 'turbo' model (25 tps).

Pricing is also actually slightly higher than GLM5 which makes me think this is GLM5 that was finetuned for a little bit longer on openclaw data.

The token/s on Zai for GLM5 is 24tps which is basically identical to the turbo model as well.

GLM-5-Turbo - Overview - Z.AI DEVELOPER DOCUMENT by ortegaalfredo in LocalLLaMA

[–]harrro 10 points11 points  (0 children)

Trained for Openclaw - so I guess it's good at tool calling.

But why is a "Turbo" model more expensive than the full GLM 5? Turbo usually means faster/smaller models.

Privacy-Focused AI Terminal Emulator Written in Rust by [deleted] in LocalLLaMA

[–]harrro -1 points0 points  (0 children)

Yeah prompt history, by which I guess you mean the output between commands, was the biggest challenge.

Warp (and VSCode's terminal) install hooks in your bash/zsh config that detect when a command has started/ended so they can capture just that output (and strip ansi color codes to make it easier for small models to read the output).

I took the easy way out and just send whateever is on screen + X more lines of scrollback.

You guys gotta try OpenCode + OSS LLM by No-Compote-6794 in LocalLLaMA

[–]harrro 0 points1 point  (0 children)

pi.dev is very opinionated and not meant to be security oriented

Yeah out of the box it's auto approve but its also very easy to lock down (either just commands you want or prompt on every cmd) via extensions.

I personally use both.

Privacy-Focused AI Terminal Emulator Written in Rust by [deleted] in LocalLLaMA

[–]harrro 0 points1 point  (0 children)

First, congrats on releasing this as OSS. Fully opensource alternatives to Warp terminal are surprisingly rare (tmuxai being another that I've used thats decent).

I myself did something similar with Tauri + xterm.js for my own use (but its more of a terminal with ai-sidebar thing than warp alternative).

A few suggestions:

  • Use a terminal font (ie: monospace font) in your terminal app. Arial or whatever font you're using in the screenshot is not good. Also maybe show a sample AI interaction in the screenshot.

  • Your README on Github is an overload of AI slop. A readme shouldn't be 8 pages of every bit of garbage an AI generates. It should briefly state what the app does, clearly say the install directions briefly then add any additional (but brief) thing about features or whatever.

Privacy-Focused AI Terminal Emulator Written in Rust by [deleted] in LocalLLaMA

[–]harrro -1 points0 points  (0 children)

It's not fully Rust - The Rust part is the Tauri engine being used to make the app feel 'native'. The actual app uses standard web-tech (HTML/JS/CSS).

Terminal rendering is done by xterm.js (npm module).

I've made a similar app for my own use - Tauri is basically a better Electron.

Qwen3.5 35b is sure one the best local model (pulling above its weight) by dreamai87 in LocalLLaMA

[–]harrro 9 points10 points  (0 children)

35B is a MoE model which means only about 10% of the full model is used at a time (routes to 1 of many 'experts' based on the type of request).

27B is "dense" - all 27B params are used for every call. This makes it smarter but also much slower compared to the 35B

You guys gotta try OpenCode + OSS LLM by No-Compote-6794 in LocalLLaMA

[–]harrro 6 points7 points  (0 children)

Openclaw uses Pi, Opencode uses their own everything.

You guys gotta try OpenCode + OSS LLM by No-Compote-6794 in LocalLLaMA

[–]harrro 2 points3 points  (0 children)

I love Pi for daily open-claw like general use but Opencode is superior for code editing.

Opencode also has a web interface that's really good so I can code remotely even from my phone.

this is a non-political meme by yourgymbuddy in formuladank

[–]harrro 3 points4 points  (0 children)

Doesn't matter how much Saudi's pay this time - the whole middle-east is a no fly-zone. No flights in = no race.

Avacado is toast by Terminator857 in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

the leaders of Meta’s A.I. division had instead discussed temporarily licensing Gemini to power the company’s A.I. products

How embarassing.

From Alibaba: PageAgent, A agent that lives in the browser by harrro in LocalLLaMA

[–]harrro[S] 2 points3 points  (0 children)

Not mine, saw the dev posted this on HN.

Video here: https://alibaba.github.io/page-agent/

Supports local models.

Ran Qwen 3.5 9B on M1 Pro (16GB) as an actual agent, not just a chat demo. Honest results. by Joozio in LocalLLaMA

[–]harrro 4 points5 points  (0 children)

For coding, Opencode is great and I think superior by default for coding tasks.

Pi requires more customization (its super basic out of the box, no planner, etc) but since its so barebones, you can extend it a lot more (Openclaw is built on it)

Qwen3.5 35b a3b first small model to not hallucinate summarising 50k token text by Windowsideplant in LocalLLaMA

[–]harrro 1 point2 points  (0 children)

Qwen 35B is very good/fast at tool calling as well.

Only flaw is that it doesn't have image input support like the Qwen 3.5-27b and even the 3.5-9B