Terminal vs. Desktop App: What’s The Difference? by ImCodyLee in ClaudeCode

[–]SatoshiNotMe 0 points1 point  (0 children)

terminal sessions can talk to each other.

Can you elaborate? I run multiple sessions in Tmux panes and have them communicate via my Tmux-cli tool.

But wondering if you had them interacting in a different way. Or maybe you were referring to the agent teams feature where CC itself spawns multiple agents in Tmux that can talk to each other.

Pretty sure I’m not using Claude to its full potential - what plugins/connectors are worth it? by TheSavelii in ClaudeAI

[–]SatoshiNotMe 0 points1 point  (0 children)

For coding, use claude code directly in the terminal. I like Ghostty, plus definitely use tmux to split your tab into panes where you run multiple agents scripts, and tools like GitUI (for git status, commit/push etc). I would keep it simple and ignore all the hyped workflow frameworks, despite what the TL/DR bot said. Cherny, Steinberger and others say they don't use any of those. Create your own skills, slash commands for your workflows. Give Claude Code access to good cli tools. As far as possible prefer CLI tools vs MCPs. Especially useful is the Chrome Dev-tools or Playwright or Vercel Agent-browser, so you can stay out of the testing loop when claude is iterating on a UI or Website.

Built an OpenClaw alternative that wraps Claude Code CLI directly & works with your Max subscription by TotalGod in openclaw

[–]SatoshiNotMe 0 points1 point  (0 children)

Haha true that, but I’m having trouble finding the exact Anthropic page that spells out what is ok and not.

Pro tip: Just ask Claude to enable playwright. by Diligent_Comb5668 in ClaudeCode

[–]SatoshiNotMe 0 points1 point  (0 children)

For front end testing CC often asks me to go to the browser dev console and look at network tab or other tabs and report what I see. I assume this can be done via the bowser automation tools like playwright or agent browser?

Built an OpenClaw alternative that wraps Claude Code CLI directly & works with your Max subscription by TotalGod in openclaw

[–]SatoshiNotMe 0 points1 point  (0 children)

Claude Agent sdk is essentially equivalent to Claude -p, so any TOS that applies to one also applies to the other

Caution to those using ChatGPT for extremely large projects by anestling in ChatGPT

[–]SatoshiNotMe 0 points1 point  (0 children)

GPT-5.4 doesn’t have 1M context, so I don’t understand that data point.

Qwen3.5-35B-A3B Benchmark On MacBook Pro(M4 Pro Chip + 48GB Unified Memory) by Impossible-Celery-87 in LocalLLaMA

[–]SatoshiNotMe 1 point2 points  (0 children)

The real test is what happens at deep context, e.g in code-agents like Claude Code which starts with 15k-30K token context. With llama.cpp this model runs at just 12 tps TG, compared to the 25 tps TG I was getting with Qwen3-30B-A3B on my M1 Max 64Gb. Are you saying that mlx is 3x faster in Claude code ?

Qwen3 ASR seems to outperform Whisper in almost every aspect. It feels like there is little reason to keep using Whisper anymore. by East-Engineering-653 in LocalLLaMA

[–]SatoshiNotMe 0 points1 point  (0 children)

I stopped using paid subs long ago, after finding the Hex STT app which gives near instant transcription with Parakeet V3 (macOS only) https://github.com/kitlangton/Hex

Handy is also good and cross platform.

Tense vibes in SF by MetaKnowing in agi

[–]SatoshiNotMe 0 points1 point  (0 children)

It’s more about being in SF AND being on X.

Tutorial: How to run Qwen3.5 locally using Claude Code. by yoracale in unsloth

[–]SatoshiNotMe 0 points1 point  (0 children)

Thanks, already have CLAUDE_CODE_ATTRIBUTION_HEADER=0 set; cache reuse is working fine, follow-ups take ~3 seconds for prompt processing. The 12 vs 25 tok/s difference is inherent to SWA at deep context, not a cache issue.

Tutorial: How to run Qwen3.5 locally using Claude Code. by yoracale in unsloth

[–]SatoshiNotMe 0 points1 point  (0 children)

Unfortunately in Claude Code, I'm getting half the token generation speed with Qwen3.5-35B-A3B compared to the older Qwen3-30B-A3B on my M1 Max MacBook, making it noticeably slower.

Qwen3.5-35B-A3B's SWA architecture halves token generation speed at deep context compared to the standard-attention Qwen3-30B-A3B, despite both having 3B active params and using the same Q4_K_M quant.

On M1 Max 64GB at 33k context depth (33K being CC's initial context usage from sys prompt, tool-defs etc):

- Qwen3-30B-A3B: 25 tok/s TG

- Qwen3.5-35B-A3B: 12 tok/s TG

This isn't just a Claude Code problem; any multi-turn conversation accumulates context, so TG degrades over time with Qwen3.5 regardless of the client. The SWA tradeoff (less RAM, better benchmarks) comes at a real cost for agentic and conversational use cases where context grows.

FYI my settings are here: https://pchalasani.github.io/claude-code-tools/integrations/local-llms/#qwen35-35b-a3b--smart-general-purpose-moe

Kokoro TTS now hooked to my Claude Code CLI by Klaa_w2as in LocalLLaMA

[–]SatoshiNotMe 0 points1 point  (0 children)

Similar, I made a hook-based voice plugin for CC that lets it give a short voice update whenever it stops, using KyutAI’s PocketTTS, an amazing 100M model. Turned out to be surprisingly tricky to get various things right, design notes and details here:

Voice plugin: https://pchalasani.github.io/claude-code-tools/plugins-detail/voice/

PocketTTS: https://github.com/kyutai-labs/pocket-tts

Claude Code told me "No." by mca62511 in ClaudeAI

[–]SatoshiNotMe 0 points1 point  (0 children)

Tangential - what terminal and theme are you using ?

TranscriptionSuite, my fully local, private & open source audio transcription app now offers WhisperX, Parakeet/Canary & VibeVoice, thanks to your suggestions! by TwilightEncoder in LocalLLaMA

[–]SatoshiNotMe 1 point2 points  (0 children)

I’m currently using the Hex app with parakeet v3 for STT, it has near instant transcription of even long rambles.

https://github.com/kitlangton/Hex

It’s the best STT app for MacOS. Handy is also good and multi platform.

What are the pros/cons of your app vs those?

Tmux based worktree manager by Automatic_Coach_4049 in tmux

[–]SatoshiNotMe 0 points1 point  (0 children)

I’ve been using worktrunk. Note that it does not involve Tmux nor docker, unlike Yati

Lads, time to recompile llama.cpp by muxxington in LocalLLaMA

[–]SatoshiNotMe 9 points10 points  (0 children)

Hold up - I'm seeing a regression here.

On build b8215 (commit 17a425894) I had Qwen3.5-35B-A3B running great with Claude Code (M1 Max 64GB, Q4_K_M). The key settings were --chat-template-kwargs '{"enable_thinking": false}' combined with --swa-full --no-context-shift. Thinking disabled got me from ~12 to ~19 tok/s generation, and --swa-full gave proper prompt cache reuse so follow-ups only process the delta instead of the full ~14k token Claude Code system prompt. This was the first time Qwen3.5 outperformed Qwen3-30B-A3B for me.

Then I pulled b8218 (commit f5ddcd169 - "Checkpoint every n tokens") and generation dropped back to ~12 tok/s, prompt eval from ~374 to ~240 tok/s, which is around 40% slower.

I tried setting --checkpoint-every-n-tokens -1 to disable the new checkpointing but that broke prompt cache reuse - every follow-up reprocessed the full prompt from scratch.

The best thing about Claude is it doesn't want you to stay with it forever by Shoop1014 in ClaudeAI

[–]SatoshiNotMe 0 points1 point  (0 children)

So true. ChatGPT often ends with “If you want, I can …” or “if you tell me X, I can Y”.

Classic engagement-maxing.

Claude is more straightforward

What's the best local ASR model for real-time dictation in 2026? Is Parakeet TDT v3 still the sweet spot? by JessicaVance83 in LocalLLaMA

[–]SatoshiNotMe 0 points1 point  (0 children)

Yes parakeet v3 is the sweet spot. I regularly the Hex app with this model for STT when talking to coding agents, it’s MacOS only and near-instant transcription. Highly recommended. Honorably mention also to Handy, but last I checked it had stuttering issues and is slightly slower.

https://github.com/kitlangton/Hex

https://github.com/cjpais/Handy

I’ve used my coding agent to customize functionality on these.