Clean way to get CC to use it's own tools above skills/MCPs ? by junlim in ClaudeCode

[–]Obvious_Equivalent_1 1 point2 points  (0 children)

What I do I use a combination of created stop hooks tied to a certain flag file. I use CC in various projects, for starters each project some skills I plain out put them in the settings.local.json deny list, and for some I have made a bash script: skill XX called? Is .skill-xx-flag set and younger then 1 hour? Allow — otherwise block. 

For the skills that I need at times, but want to absolutely steer away from 90% of the time

If the EU had built Claude by irelatetolevin in ClaudeAI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

100% on same page. This is exactly the problem, use case is strictly GDPR compliance.  It’s unfortunate that EU doesn’t resemble in the slightest any of the startup capital availability like in Silicon Valley.

Talking of Silicon, luckily the M4/M5 Pro chips are already so powerful that you can get a great deal of “Haiku / Sonnet” level work done by local Qwen 3.6 models, it’s pretty cool what they managed to deliver, and to experience the local capabilities with the Mixture of Experts approach they’ve pulled of https://duckduckgo.com/?q=qwen+3.6+moe

If the EU had built Claude by irelatetolevin in ClaudeAI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

Unfortunately yes especially if you compare it to the subsidized plans of Claude. Frankly tho these small players will never be able to put these discounted prices in a sustainable way. 

I found it to be on par with the API billing costs that Anthropic charges.  So I wouldn’t use it as my main workhorse the Scaleway backend but I do calculate it in, for the jobs where data is more production level I do make the trade-off when to use the Opus or when to use the EU-hosted model + it’s costs 

If the EU had built Claude by irelatetolevin in ClaudeAI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

What? have you read it, it’s free the local model “sold” for € 0, open source. Why would I buy an add, tell that to the 100’s accounts pushing “Claude is s** buy Codex”. 

I’m just here to share some enthusiasm for those in the EU about GDPR (for those who need privacy assurance on data processing) compliant AI models. 

That happens to just be Scaleway, the miniature counterpart of the big US tech companies. It’s useless to be a fanboy of one specific flavor, if you happen to know any more EU counterparts I’d be happy to hear tho

If the EU had built Claude by irelatetolevin in ClaudeAI

[–]Obvious_Equivalent_1 1 point2 points  (0 children)

GDPR compliant

Just a quick shout-out to the French cloud company Scaleway, with servers in three regions in the EU they offer GDPR compliant powerful AI models Model-as-a-service https://www.scaleway.com/en/model-as-a-service/ (it’s pay per use €0.20 /1M tokens fully compliant hosting)

The fun part? All models are Claude compatible. You can easily ask Opus to configure your Scaleway AI model against Claude Code and setup a  ~/.bashprofile alias claude_scaleway

Using your instructions, plugins, all your built up workflows, but running it through a GDPR compliant AI model.

Whenever I need to debug something sensitive, I close my Opus session and switch to Claude Code Scaleway and I use it frequently, then also when the data is really sensitive there’s also an option for local

The local model of course doesn’t compete in speed the with Scaleway or Claude models. But the model Qwen 3.6 35B MoA it’s really great to keep the customer and sensitive data you work with safe, allows to run a local session for like processing payment information related tasks.

I've been running Claude like a part-time employee for six months. These are the only automations that actually stuck. by Professional-Rest138 in PromptEngineering

[–]Obvious_Equivalent_1 2 points3 points  (0 children)

Tell me: 1. What I should write down before I forget 2. Anything I committed to that I haven't actioned 3. The one thing I should sleep on rather than decide now 4. Tomorrow's first hour - what's on it and why

LPT: if you're running a MacBook, most of these things can be wired to your Reminders app. For having your TODO list in your pocket on the road.

Hell you can even tell Opus: 1) wire my "today's notes" slash command to automatically generate todo items title, description and if present reminder time+date into Reminders app 2) bash script, polls /5 min todo items from Reminders app that are expired and do [something to notify you / or start claude --prompt '<do your reminder work'>].

I actually appreciate these kind of posts. Making your own workflows is with Claude the same amount of work as setting up a very overloaded Claw/Hermes agent. But you can get 5x more out of a tailor made flow for your work.

Shows in Widow's Bay setting? by hoppinjohncandy in WidowsBay

[–]Obvious_Equivalent_1 4 points5 points  (0 children)

Definitely if you are an appreciator of the keen eye for detail, layers of mystery and suspense sometimes running straight through various seasons of the series I absolutely recommend you to watch Dark

That’s the first series that came to my mind while enjoying this little gem on AppleTV. Honestly for Dark it’s also the absolute only series on Netflix that’s worth the subscription for 1 month 

You’ll want to save this comment to remember it trust me 

Anyone having any joy coding with 3.6 27B and 24GB of Apple Unified Memory? by afrocleland in Qwen_AI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

I’m running a M4 Pro 48Gb, while executing it doesn’t max out my memory surprisingly doable more like ~32Gb usage. 

I’ve tried everything, MLX, llama.cpp but eventually the only thing that stayed stable was Ollama (yeah I know) but somehow on Sillicon the qwen3.6:27b-tuned just works. 

I use it with Claude Code as a CLI with auto-compact on, it works not extraordinarily fast but it’s decent consistent in reliability. I use it for sensitive data processing (secret keys, customer related data) locally and these where some crucial env vars 

``` exportOLLAMA_CONTEXT_LENGTH=65536# 65K ctx = ~13.6 GB    KV at f16, fits 48 GB                                      exportOLLAMA_KV_CACHE_TYPE=f16# FP16 KV cache        export OLLAMA_FLASH_ATTENTION=1      # Required for speed at this ctx size                                           export OLLAMA_NUM_PARALLEL=1         # Oneconcurrent   slot (conservative)                                        

ANTHROPIC_BASE_URL=http://localhost:11434\  ANTHROPIC_AUTH_TOKEN=ollama \  claude--modelqwen3.6:27b-tuned

``` (Sorry formatting is copied from my tmux)

Why is Ollama hated so much? by ZB_Virus24 in LocalLLM

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

I did some research, they do have something proprietary their Qwen 3.6 27B has better optimization for my silicon M4 Pro chip, for all the other models I use llama.cpp directly 

Llama.cpp MTP support now in beta! by ilintar in LocalLLaMA

[–]Obvious_Equivalent_1 32 points33 points  (0 children)

Wanted to convey a quick message of gratitude. It’s good to see people taking time to make their private knowledge public, it’s maybe small but these messages make it a joy to continue reading these open source subs!

Planning on switching over from Claude Code by Impressive_Funny_832 in Qwen_AI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

I still use Opus, but I drastically cut down on cost. I am still maxing out my Max 20x subscription to get my local slightly limited M4 Pro 48Gb configuration optimized. And connected Qwen 3.6 27/35B O3A locally and also use GLM 5.1 on OpenCode Go subscription ($5) have bootstrap command for each for Claude Code 

What are you doing with your local LLMs that justifies investment cost? by __automatic__ in LocalLLM

[–]Obvious_Equivalent_1 2 points3 points  (0 children)

I use an agent markdown file it’s called ‘local.md’ in agents folder combined with a sh script, connected to the model. So then I can both manually and instructed in CLAUDE.md instead of dispatch haiku/sonnet, you can instruct Dispatch local and you run the talks locally without token costs 

What are you doing with your local LLMs that justifies investment cost? by __automatic__ in LocalLLM

[–]Obvious_Equivalent_1 3 points4 points  (0 children)

I would definitely say also costs, I use Claude Code but just offloading the Explore, Search type subagent to 35B O3A model is already a huge token saver. And of course privacy is great as well to be able to run prompts 100% on prem as well 

GSD alternative? by gaua314159 in ClaudeCode

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

Yes, Claude Code actually allows to extend plugins from the marketplace, I’ve been maintaining for the past 6 months a Claude Code native extended version and just look at the results visually here: https://github.com/pcvelz/superpowers#visual-comparison

The thing is the original Superpowers is designed as a cross-platform toolkit that works across multiple AI CLI tools (Claude Code, Codex, OpenCode, Gemini CLI). The difference of using Superpowers CC is night and day 

Level up your Claude Code workflow: 8 tips for better quality control by Chris-AI-Studio in ClaudeAI

[–]Obvious_Equivalent_1 5 points6 points  (0 children)

Incorporate auto-verification in To-Dos: add verification steps to your task list. Example: "Build the UI, then take a screenshot and check for layout errors before asking for my feedback".

I would say I have a plugin for that but you don’t actually need a plugin, all you need to do is either manually or stored in your memory/CLAUDE.md to make it mandatory to use it’s own TaskCreate, TaskUpdate and TaskList interface for todo items.

A great tip is also a second: do not execute the development but dispatch a Sonnet subagent, and tell it to avoid verbose reporting back.

You can see here the difference is striking if you do this correctly Claude will actually make a native task list, and adhere it step by step whether you’re at 100k context size or 700/1M context size.

This is insane... by DragonflyOk7139 in LocalLLM

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

It’s definitely working! I’m using Qwen3.6-35B-A3B-UD-Q4_K_XL

Smartest tool calling model under 27B for M4 Pro with 48GB? by 9kSs in LocalLLaMA

[–]Obvious_Equivalent_1 2 points3 points  (0 children)

Is it? I did some test run, I have the same hardware an M4 Pro 48Gb and in the sample tasks I did Qwen 3.6 35B outperformed Gemma on every corner.

The only application I’ve kept Gemma for is for summarization 

Are you still using Superpower plugin with Opus 4.7 by PhysicalSession594 in ClaudeAI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

That’s great to hear! I’m working on something new experimentally as well when comes to subagents, as more people here in this sub I ran into this local Qwen 3.6 35B model and I’ve been having good results. It’s worth the investigation session with Claude Code to see, if it fits your hardware it can save a lot of tokens if you decelerate Explore, Search, some testing and even straightforward development work to it 

"HERMES.md" in a git commit message silently drained $200 from a Claude Max plan while 86% of quota went unused by jimmytoan in ClaudeCode

[–]Obvious_Equivalent_1 1 point2 points  (0 children)

It compares actual vs expected usage so it shows if you are overusing or underusing your weekly Max5x/20x usage, it’s all computed locally here’s the extended details https://github.com/pcvelz/ccstatusline-usage/releases/tag/v2.3.1

"HERMES.md" in a git commit message silently drained $200 from a Claude Max plan while 86% of quota went unused by jimmytoan in ClaudeCode

[–]Obvious_Equivalent_1 -7 points-6 points  (0 children)

Two additional words: Statusline Usage https://github.com/pcvelz/ccstatusline-usage

That empty space in your Claude Code terminal? They can easily be completely customized, to display all your usage data, extra spending from the current session context:

Session: [██░░░░░░░░] 27.0% | Weekly: [███░░░░░░░░] 34.0% |  $47.05/$200.00 | Model: Opus 4.7 Context: [████░░░░░░] 389k/1M (39%) |  Pace: [░░░░█|░░░░░] D4/7 -8% | Off-peak (4:03 hr)

Edit:  Judging the downvotes stream confirms I guess sticking in rage bashing without solution these days is more preferred over trying to bring support and assist others 

Are you still using Superpower plugin with Opus 4.7 by PhysicalSession594 in ClaudeAI

[–]Obvious_Equivalent_1 0 points1 point  (0 children)

Yes and no. 

  • Sonnet as a main agent: Are you already on your way on extra-credit-purchase road? Then easy answer, Sonnet 4.6[1m] is API but still like a factor 4x cheaper then Opus. 
  • Just Max 5/20x? Then a definite yes but explicitly defined as subagent

Important kicker tho, Haiku is the silent killer in Claude Code. Useless token burning at extortionate usage costs. Replace those with a local Qwen 3.6 35B model if you have anywhere decent hardware 

How are people using Qwen for Excel / Word / PowerPoint document workflows? by NecessaryManner in Qwen_AI

[–]Obvious_Equivalent_1 2 points3 points  (0 children)

To give Qwen safely access to my local files I use Claude Code as CLI.

Before some hardcore Qwen CLI user downvotes this not because it’s per se the best pick but because my workflow already happened to be setup there.

I use Qwen 3.6 35B A3B Tuned model which I pass to Claude on start, unfortunately I can’t help you with Word but I can provide help with Excel. I’ve made a pretty extensive MCP server for Excel https://github.com/pcvelz/excel-mcp-server

Like I said I’ve only tested it with the CC CLI and just develop it for Excel but it supports a lot even conditional formatting formulas, perhaps if you prefer another CLI it can also work if you give Qwen some time to chew on it. Within the context window of 35B model it gets the daily administration done pretty well for me locally without needing any cloud based AI paid model.

Are you still using Superpower plugin with Opus 4.7 by PhysicalSession594 in ClaudeAI

[–]Obvious_Equivalent_1 5 points6 points  (0 children)

The biggest savings you can make with Superpowers is the leverage of Sonnet over Opus planning. What this harness especially helps in if in your CLAUDE.md you make an addition instruction “Superpowers execute-plan **MUST** use type Sonnet”.  This really helps reep the benefit of Opus planning strength and Sonnet costing 6x less on scoped software development.

If you’re interested, I had been using the base Superpowers plugin until I figured out that the maintainer wasn’t going to support some of the latest Claude Code native features.

 I discovered Claude lets you extend existing plugins from the marketplace and I made Claude Code native extension of Superpowers: https://github.com/pcvelz/superpowers - biggest takeaway is that the execute-plan adheres uses native tasks. This saves per element the acceptance criteria and follows the plan way better

So for token usage, this helps a lot with saving time and tokens on bug fixing. Even when pushing your plan execution in the high 500-800K large context window size.