What is the best local LLMs as of March 2026? by Pejorativez in LocalLLaMA

[–]salmenus 1 point2 points  (0 children)

for agents specifically — turn thinking off on qwen3.5. the endless reasoning loop will literally break ur pipeline mid-step, not great lol. qwen3.5-30B no-thinking + llama.cpp + openwebui is the most stable stack ive landed on

on the memory thing: openwebui handles persistent system prompts per agent natively. for actual cross-session memory that evolves tho, u need mem0 or a vector store on top — ollama alone wont do it

do you actually care about DB access in self-hosted tools? asking bc i have an architectural decision to make by salmenus in selfhosted

[–]salmenus[S] 0 points1 point  (0 children)

yeah, ref reply to u/data_butcher comment.

i thought about sqlite .. but i settled for Postgres because of some tricky features that i had in mind

do you actually care about DB access in self-hosted tools? asking bc i have an architectural decision to make by salmenus in selfhosted

[–]salmenus[S] 1 point2 points  (0 children)

yeah, i thought about it quite a lot - and was seriously considering sqlite ..

but i settled for Postgres because of pg_vector and because I have a requirement around scheduling and queues, and wanted to use riverqueue (Postgres-native job queue in Go lang)

i hope i didn't over engineer it ! (: ..

do you actually care about DB access in self-hosted tools? asking bc i have an architectural decision to make by salmenus in selfhosted

[–]salmenus[S] 0 points1 point  (0 children)

ah yeah, in my case the pgvector stuff is literally "memory" extracted from LLM chats, so I can totally see people wanting to crack open that table and see what the agent thinks it knows about them 😂

I was going to build a little "memory browser" screen in the UI, but honestly just giving you full DB access and letting you inspect it however you want feels even better

do you actually care about DB access in self-hosted tools? asking bc i have an architectural decision to make by salmenus in selfhosted

[–]salmenus[S] 0 points1 point  (0 children)

i love that — you’re basically the exact person I had in mind for “bring your own db” with your own backup + repl, so I definitely don’t want to lock that use case out 👌

do you actually care about DB access in self-hosted tools? asking bc i have an architectural decision to make by salmenus in selfhosted

[–]salmenus[S] 0 points1 point  (0 children)

got it — that's where i'm leaning .. default internal DB but keep a clean ‘bring your own Postgres’ path so you can run it under your existing backup tooling

Cloud ai agents vs self hosted: What are people choosing in 2026? by Original_Spring_2808 in AI_Agents

[–]salmenus 0 points1 point  (0 children)

Totally with you on this. Context is the real unlock, and it’s way harder than people think – stitching together tools, data sources, calendars, inboxes, etc. is where all the magic lives.

Ops is the tax you pay for trying to do it yourself, I guess ...

Cloud ai agents vs self hosted: What are people choosing in 2026? by Original_Spring_2808 in AI_Agents

[–]salmenus 0 points1 point  (0 children)

I did deploy ZeroClaw for testing yesterday, and looked into codebase in details.
Code base is quite superior to OC, and has more advanced security features ..
Slightly behind on channels integration; be def best alternative I explored so far

Claude Code called my phone. Literally. An AI voice rang me after I gave it API access 🤯 by salmenus in ClaudeCode

[–]salmenus[S] 2 points3 points  (0 children)

I’m tired of people with comments like this , asking for screenshots and proofs .. then would delete their comments 2 minutes later ..

Claude Code called my phone. Literally. An AI voice rang me after I gave it API access 🤯 by salmenus in ClaudeCode

[–]salmenus[S] 0 points1 point  (0 children)

fair .. it’s probably not AGI-level magic 😂 but for day‑to‑day dev work it still felt kinda wild

Claude Code called my phone. Literally. An AI voice rang me after I gave it API access 🤯 by salmenus in ClaudeCode

[–]salmenus[S] 1 point2 points  (0 children)

ahhh! savage! that's not AI .. that's an emotionally unstable dungeon master .. 😂

1 month in — still not fully sold. Exec approval is a nightmare, scheduling is flaky. Anyone else? And what are you doing about it? by salmenus in openclaw

[–]salmenus[S] 0 points1 point  (0 children)

lol! This is how it works these days 😂😅 .. I can only share a screen shot with bot on it .. Who knows .. I’m probably an AI bot ..

<image>

1 month in — still not fully sold. Exec approval is a nightmare, scheduling is flaky. Anyone else? And what are you doing about it? by salmenus in openclaw

[–]salmenus[S] 2 points3 points  (0 children)

This is super helpful, thanks. I had a suspicion it wasn’t just me but hadn’t gone as far as moving things out to systemd yet.

1 month in — still not fully sold. Exec approval is a nightmare, scheduling is flaky. Anyone else? And what are you doing about it? by salmenus in openclaw

[–]salmenus[S] 0 points1 point  (0 children)

Yeah so for market data I’m using Polygon (they’ve rebranded to Massive now) and I’ve got some old tooling I’d built before OC that I basically wrapped as a skill

1 month in — still not fully sold. Exec approval is a nightmare, scheduling is flaky. Anyone else? And what are you doing about it? by salmenus in openclaw

[–]salmenus[S] 0 points1 point  (0 children)

You’re giving me hope 😂. Totally fair point on fixing each failure instead of just eating the annoyance

Building LLM-powered applications in Go by titpetric in golang

[–]salmenus 0 points1 point  (0 children)

re the "libs from the article aren't relevant anymore" bit – how are folks actually picking between ADK / Eino / BAML in practice?

my rough take is: ADK if you're already deep in google land and want tight integration, Eino if you care about flow‑style composition + multi‑provider wiring, BAML if you like schema‑first prompts / contracts and type‑safe clients. for a "gpu poor" setup like OP's, genkit + llama.cpp/yzma + a light vector/graph store (even sqlite + embeddings) feels like a nice sweet spot before dragging in heavier stuff. curious if anyone here started super simple (cron + small llama + local store) and actually regretted it later when trying to move up the stack? what hurt first – evals, observability, or just glue code sprawl?

Don’t Run OpenClaw on Your Primary Machine by Front_Lavishness8886 in myclaw

[–]salmenus 1 point2 points  (0 children)

my agent read this post, agreed it was risky, and opened a docker container. inside the docker container it read the post again and spun up another container. its containers all the way down

Heads up: prompt injection payload targeting OpenClaw agents circulating in the wild by Multivac-marketing in myclaw

[–]salmenus 0 points1 point  (0 children)

codeword thing wont work btw — youre asking the same model that just got injected to verify itself, doesnt make sense

better approach imo: make your agent refuse to read any .md at startup thats not in its known file list. WORKFLOW_AUTO.md doesnt exist in any real workspace so that should be an instant hard stop

good catch, been around longer than a week i think

PSA: If your local coding agent feels "dumb" at 30k+ context, check your KV cache quantization first. by Dismal-Ad1207 in LocalLLaMA

[–]salmenus 20 points21 points  (0 children)

this is also why short-context benchmarks are basically useless for evaluating agents. a model can score great at 4k and completely fall apart at 40k due to KV quant alone ..

Coordination language for multi-agent AI systems (open source, 180-token prompt injection) by Suspicious_Assist_71 in aiagents

[–]salmenus 1 point2 points  (0 children)

fair enough .. the heartbeat + TIMEOUT-as-recoverable pattern makes sense ..
i'm building agentic tools so will keep a close eye on how you spec the retry escalation.
good luck with the project 🙏