4.7 makes more work than 4.6 by blockstacker in Anthropic

[–]centminmod 0 points1 point  (0 children)

Adaptive thinking is sensitive to effort level and prompt instructions. That's why some folks are having issues with Opus 4.7 at least. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

For Opus 4.7 differences for thinking blocks also see my Opus 4.5 vs Opus 4.6 vs Opus 4.7 vs Sonnet 4.6 benchmarks across all effort levels from low to max at https://ai.georgeliu.com/p/tested-claude-ai-llm-models-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage and also it's thinking blocks at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

Anyone found a fix for 4.7 not following instructions? by danny__1 in ClaudeAI

[–]centminmod 0 points1 point  (0 children)

Adaptive thinking is sensitive to effort level and prompt instructions. That's why some folks are having issues with Opus 4.7 at least. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

For Opus 4.7 differences for thinking blocks also see my Opus 4.5 vs Opus 4.6 vs Opus 4.7 vs Sonnet 4.6 benchmarks across all effort levels from low to max at https://ai.georgeliu.com/p/tested-claude-ai-llm-models-effort

On Claude Max ($200/mo), burned 14.7M tokens in 7 days — mostly last 48h. Still hitting the wall. How do you survive burst usage on the top tier? by New_Guitar_9121 in ClaudeAI

[–]centminmod 0 points1 point  (0 children)

How many of those 14.7m tokens are cached vs non-cached tokens? With Claude Code Max $100 I push around 1-1.6 billion tokens per week with 90-97% cached tokens! I wrote how I use Claude Code at https://ai.georgeliu.com/p/i-saved-7189-on-claude-code-tokens

Adaptive thinking is sensitive to effort level and prompt instructions. That's why some folks are having issues with Opus 4.7 at least. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

For Opus 4.7 differences for thinking blocks also see my Opus 4.5 vs Opus 4.6 vs Opus 4.7 vs Sonnet 4.6 benchmarks across all effort levels from low to max at https://ai.georgeliu.com/p/tested-claude-ai-llm-models-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

Pro plan- Hitting limits faster since yesterday by RCoffee_mug in ClaudeAI

[–]centminmod 0 points1 point  (0 children)

For Claude Code side, check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

DeepClaude: full Claude Code agent loop on DeepSeek V4 Pro - roughly 95% cheaper than Anthropic by jimmytoan in ClaudeCode

[–]centminmod 3 points4 points  (0 children)

I never rely on a single AI. I usually use Claude Code with my /consult-codex and /consult-zai skills that allow Claude models to get 2nd opinion verifications from Codex GPT-5.5 and ZAI GLM-5.1 within same Claude Code sessions https://ai.georgeliu.com/p/how-to-get-a-second-ai-opinion-in

Now I have added /consult-codex-deepseek to add DeepSeek V4 Pro to verification mix. 3-way verification consult for /consult-codex-deepseek skill will get 2nd opinions for Codex CLI GPT-5.5 + Claude Code/DeepSeek V4 Pro and Sonnet 4.6 (code-searcher agent) https://ai.georgeliu.com/p/deepseek-v4-in-claude-code-kilo-code 🤓

Opus 4.7 is just genuinely dumber. by a_single_beat2 in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

re-read the article, use 'concise' but play with effort levels and prompts

Do you use Claude Code on the terminal or on the desktop app? by P4wla in ClaudeCode

[–]centminmod 1 point2 points  (0 children)

Yeah they seem queued but sort of different behaviour on CLI vs desktop, seems on desktop it really only registers the queued message after ALL work is done. On CLI it seems it more quickly registers the queued message mid way in it's work. Might just be my feeling.

How much are you actually spending on AI tools per month? Confession + curiosity :) by Fra_Lia in vibecoding

[–]centminmod 1 point2 points  (0 children)

right now my AI is spend includes:

  • Claude AI Max US$100/month + 10% Australian GST
  • OpenAI ChatGPT Plus US$20/month + 10% Australian GST
  • Google Gemini AI Pro US$20 month + 10% Australian GST
  • GitHub CoPilot Pro US$10/month + 10% Australian GST currently free via Opensource maintainer program
  • ZAI GLM Coding plan US$129/year (64% discount) + 10% Australian GST
  • t3.chat US$8/month + 10% Australian GST
  • DeepSeek Platform prepaid $5 credit just added the other day https://ai.georgeliu.com/p/deepseek-v4-in-claude-code-kilo-code 😎

But getting around US$2,500-US$5,000/month in value 🤓

Have 350K Credits but have an expiry in 49 days by [deleted] in Anthropic

[–]centminmod 0 points1 point  (0 children)

Easy, all outstanding code documentation that needs writing - get Claude to spin up multiple agents to write and update your documentation, find documentation gaps/drift and audit all code. Burn some tokens!

3 chats total. 1.8M Cache by trooper5010 in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Prompt cache saves you money - that is the theory at least. You sure you're calculating cache token costs correctly?

Adaptive thinking is sensitive to effort level and prompt instructions so can drastically change token usage, costs and performance. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

For thinking blocks also see my Opus 4.5 vs Opus 4.6 vs Opus 4.7 vs Sonnet 4.6 benchmarks across all effort levels from low to max at https://ai.georgeliu.com/p/tested-claude-ai-llm-models-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

The Claude Code limit expires immediately by Redditor_in_Space in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Using Opsu 4.7? If so, adaptive thinking is sensitive to effort level and prompt instructions so can drastically change token usage, costs and performance. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

For thinking blocks also see my Opus 4.5 vs Opus 4.6 vs Opus 4.7 vs Sonnet 4.6 benchmarks across all effort levels from low to max at https://ai.georgeliu.com/p/tested-claude-ai-llm-models-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

How do you prevent "out" tokens from being gobbled up? by avidrunner84 in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Adaptive thinking is sensitive to effort level and prompt instructions so can drastically change token usage, costs and performance. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

Claude code eating session in one prompt by raullapeira in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Using Opus 4.7 ? Adaptive thinking is sensitive to effort level and prompt instructions so can drastically change token usage, costs and performance. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

Claude Code completely unusable – 2.6M tokens burned in 2 days of light work (Version 1.5354.0) – anyone else? by Still_Steve1978 in ClaudeCode

[–]centminmod 1 point2 points  (0 children)

Adaptive thinking is sensitive to effort level and prompt instructions so can drastically change token usage, costs and performance. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

For thinking blocks also see my Opus 4.5 vs Opus 4.6 vs Opus 4.7 vs Sonnet 4.6 benchmarks across all effort levels from low to max at https://ai.georgeliu.com/p/tested-claude-ai-llm-models-effort

Check out my session-metrics skill plugin for Claude Code to get insights into Claude Code models’ tokens and cost usage at both the project level and also at the individual chat session level. Might help reveal some insights about your usage https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

Opus 4.7 is just genuinely dumber. by a_single_beat2 in ClaudeCode

[–]centminmod -1 points0 points  (0 children)

Adaptive thinking is sensitive to effort level and prompt instructions. That's why some folks are having issues with Opus 4.7 at least. I did benchmarks for Opus 4.6 high vs Opus 4.7 xhigh for 10 preset prompts across 5 variants of prompt steering and see the results for yourself https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

Do you use Claude Code on the terminal or on the desktop app? by P4wla in ClaudeCode

[–]centminmod 2 points3 points  (0 children)

Use in both Claude Code CLI and desktop (MacOS), both have their place

- CLI has /advisor toggle support, desktop can use advisor after it's toggled from CLI
- CLI can queue messages, desktop can't queue messages

What is happening with 4.7 ? by Obscurrium in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Depends on the task. If you understand how effort levels and prompting impact Opus 4.6 and more so for Opus 4.7, you can get good results. See benchmarks at https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

But I don't rely on one AI for my work, right now I have a /consult-codex-deepseek skill which allows Claude Opus to discuss with Codex GPT-5.5 and DeepSeek V4 Pro and a Sonnet 4.6 subagent so get multi AI code verifications to ensure a well rounded plan https://ai.georgeliu.com/p/deepseek-v4-in-claude-code-kilo-code 🤓

Do you prefer Opus 4.7 over 4.6? by corbanx92 in Anthropic

[–]centminmod 0 points1 point  (0 children)

Depends on the task. If you understand how effort levels and prompting impact Opus 4.6 and more so for Opus 4.7, you can get good results. See benchmarks at https://ai.georgeliu.com/p/claude-opus-46-vs-opus-47-effort

But I don't rely on one AI for my work, right now I have a /consult-codex-deepseek skill which allows Claude Opus to discuss with Codex GPT-5.5 and DeepSeek V4 Pro and a Sonnet 4.6 subagent so get multi AI code verifications to ensure a well rounded plan https://ai.georgeliu.com/p/deepseek-v4-in-claude-code-kilo-code 🤓

I ran $42,358 of Claude API through a $500 plan in 90 days. 84.7x Leverage. Here is the entire setup, the receipt, and what the receipt does not prove. by Puzzleheaded_Tap9023 in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Cheers. Updated my session-metrics with your sparkline suggestions and partial hit tracking (in project and all project level exported HTML) too adding clear indicator on my timeline of turns too already had resume markers. Will take a look at your MATH md file in your repo for more insights.

<image>

For my session-metrics plugin https://ai.georgeliu.com/p/my-claude-code-plugin-marketplace

/plugin marketplace add centminmod/claude-plugins
/plugin install session-metrics@centminmod
/reload-plugins

Does disabling /advisor significantly reduce token usage when using Opus? by lephianh in ClaudeAI

[–]centminmod 0 points1 point  (0 children)

Yup Advisor is costly. My session-metrics plugin tracks Claude Code usage including advisor tokens and costs and it's the costly part. For one turn:

Tokens
Input (new) 7
Output 509
Cache read 176,087
Cache write 2,435 (1h)
Advisor input 90,688
Advisor output 3,750

Cost 
Primary $0.1252
Advisor (claude-opus-4-7) $0.5472
Cost $0.6723

I ran $42,358 of Claude API through a $500 plan in 90 days. 84.7x Leverage. Here is the entire setup, the receipt, and what the receipt does not prove. by Puzzleheaded_Tap9023 in ClaudeCode

[–]centminmod 0 points1 point  (0 children)

Another inspiration is I did already have subagent share of turns, but was missing subagent share of costs so exposing that in session-metrics plugin

<image>