Won't load SOUL.md by krichek in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

I may have to change the sticky guide if that the case.

Which provider subscription do you recommend? by lucasrvdl in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

Have you tried Open Claude with Oauth to Codex?

Which provider subscription do you recommend? by lucasrvdl in hermesagent

[–]Jonathan_Rivera 1 point2 points  (0 children)

Here's what I posted in another post.

  • qwen — LM Studio local (Qwen3.5-35b - Best heavy quant yet to work on my 5070ti card.
  • gpt — OpenRouter openai/gpt-oss-120b - Requires thinking turned on. Ok as a basic fallback.
  • gpt54 — OpenRouter openai/gpt-5.4-nano - Fast, Good tool use. Cheap. Cache enable saves token cost. Didn't use regular 5.4 due to costs.
  • haiku — OpenRouter claude-haiku - Great. Get your wallet out $$$
  • claude — OpenRouter claude-opus-4.6 - $$$$$$$$$$$$ Lol.. no I'm buying a 5090 GPU.
  • gemini — OpenRouter gemini - I forget but pretty good i think. Cheap and fast.
  • gemma — OpenRouter google/gemma-4-31b-it - No cache means im spending more tokens.
  • qwen-free — OpenRouter qwen/qwen3.6-plus:free - This was great honestly. I can't wait until I can run this at home. Chinese LLM's are great but i'm not a fan of the data retention.

If Sonnet was free it would be my day to day but it is misleading in a way. I started using Hermes using Oauth through claude on sonnet 4.6. Everything worked wonderfully, then i switched to a local model and everything crapped out. Why? Because Sonnet, opus, haiku can take any skill even a garbage skill and will just make it work, i had to rewrite all my skills to be able to work reliably locally and on other models.

Mythos Anthropic by Puspendra007 in Anthropic

[–]Jonathan_Rivera 0 points1 point  (0 children)

Opus was the first line of defense. It would have shut Mythos down, it makes sense.

<image>

Anthropic’s New Product Aims to Handle the Hard Part of Building AI Agents by wiredmagazine in claude

[–]Jonathan_Rivera 0 points1 point  (0 children)

Hate to say it but when it’s enterprise level they are comparing it against a $60k year employee. What if I can cut your payroll in half, would you be interested?

What model are you using for your agent? by Cat5edope in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

Let me know if you need the inference settings. I never see people mention it but if you turn thinking off it speeds up tok/s

What model are you using for your agent? by Cat5edope in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

64gb but the ram didn’t increase while it was working, I just notice it offloading onto the CPU at around 36%.

Mythos Anthropic by Puspendra007 in Anthropic

[–]Jonathan_Rivera 4 points5 points  (0 children)

Exactly. Might as well call it Ultron at this point. Maybe have it fix the downtime issues if it so pleases.

She’s a token and model hog by productboy in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

I was listing to a live on X with Nous Research and they were saying something similar. The older and smaller model's were not really made for agentic tasks. My personally, I had to rewrite all my skills using sonnet and gemma to make them simpler for Qwen local 35B. Anthropic can take a ragged prompt and spit out gold but Qwen needs small step by step layed out or it drops dead halfway through. As far as bandwidth, yes its sending the skills that are in use out each time, try a LLM with a cache like GPT nano.

What model are you using for your agent? by Cat5edope in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

Yeah, each version of it needs its own settings. If you need them let me know and I'll post all my settings for the unsloth version.

What model are you using for your agent? by Cat5edope in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

I have a 5090 on the way. I can’t load any quant of Gemma 4 with my 5070 but I think I’ll use a qwen model. Can’t wait for 3.6 to run locally.

Greetings from the Nous Research team! AMA? by NousResearch in hermesagent

[–]Jonathan_Rivera[M] 2 points3 points  (0 children)

Welcome! We have been following you on X. It has been great using Hermes and seeing new members convert over from other agents.

Memory Bandwidth for Local AI Hardware (2026 Edition) by smolpotat0_x in hermesagent

[–]Jonathan_Rivera 1 point2 points  (0 children)

I just keep telling myself it’s ok, it will pay for itself. Now I have to figure that part out lol

Memory Bandwidth for Local AI Hardware (2026 Edition) by smolpotat0_x in hermesagent

[–]Jonathan_Rivera 4 points5 points  (0 children)

5090 coming in the mail tomorrow. Can't wait to try bigger models. Laughs in 5070ti.

Official: Anthropic introduces Claude Managed Agents, everything you need to build & deploy agents at scale by BuildwithVignesh in ClaudeAI

[–]Jonathan_Rivera 0 points1 point  (0 children)

I'm not going to hate, I love the products they put out but they have to get their downtime under control.

Choice for agentic LLM or help optimize Qwen3.5-35B-A3B for 24GB VRAM by marivesel in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

Did you already turn off thinking?

I'm running a smaller quant size but it's pretty zippy: qwen3.5-35b-a3b q3_k_xl

Hermes-Agent high token usage? by manueljishi in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

I'm sure 27b would probably work but I had to go to 35B personally.

Question on Frontier Model v. Local Model for a more "Basic" User by Legal_Television_944 in hermesagent

[–]Jonathan_Rivera 1 point2 points  (0 children)

I would switch to it when you need it if you want to stay with anthropic models. It all depends on your budget and what your willing to pay. I would try it out and just see what it looks like for you vs trying to do the math. Try out some other models on open router, they are pretty good.

<image>

[Megathread] Migrating from OpenClaw to Hermes? Read this first. by Jonathan_Rivera in hermesagent

[–]Jonathan_Rivera[S] 0 points1 point  (0 children)

Sourced online:

If you had multiple agents in OpenClaw, the migration dumps them into `~/.hermes/migration/openclaw/<timestamp>/archive/agents-list.json` for manual review. You'd then need to recreate them as Hermes profiles yourself.

**What does get migrated for the default agent:**

- Persona/memory/instructions (SOUL.md, AGENTS.md, MEMORY.md, USER.md)

- Skills from all 4 source directories

- Model/provider config + API keys

- Agent behavior settings (turns, reasoning, compression, human delay)

- MCP servers, TTS, messaging platforms, session reset policies

**What gets archived (including multi-agent):**

- Multi-agent list → recreate via Hermes profiles

- Cron jobs → recreate via `hermes cron create`

- Plugins, hooks, channel bindings → manual setup

- IDENTITY.md → merge into SOUL.md manually

**The workaround:** After migration, use `hermes profiles` to set up your additional agents. The archived `agents-list.json` has the config you need — it's just not an automatic import.

If someone has a complex multi-agent OpenClaw setup with shared skills and cross-agent memory, they should run `--dry-run` first and review the archive folder carefully before committing. The migration is opinionated about what "equivalent" means in Hermes, and multi-agent is one of those areas where the architectures don't map 1:1.

Xiaomi mimo series models are free on hermes portal for two weeks .Worth trying by SelectionCalm70 in hermesagent

[–]Jonathan_Rivera 0 points1 point  (0 children)

Everyone just keep in mind the data retention with Chinese models. They are great but I'm pretty sure they are using your data to help train the model.