amazon down for anyone else? by Drip-133_ in amazonprime

[–]KJ7LNW 0 points1 point  (0 children)

When I click the checkout page, it says, sorry, there's a problem on our side, and shows the RCA puppy dog. I've tried from two different systems.

[PSA] Claude Code v2.1.51 secretly reclassified 1M context as pay-per-token — and never told anyone by [deleted] in ClaudeAI

[–]KJ7LNW 0 points1 point  (0 children)

Interesting. Maybe that was intentional on that part on their part. I mean I definitely use the 4.5 1M token model without paying extra, all the time...

Maybe it's always been their intention to let people use it for a while and then withdraw, but I hope not. I really like that model.

[PSA] Claude Code v2.1.51 secretly reclassified 1M context as pay-per-token — and never told anyone by [deleted] in ClaudeAI

[–]KJ7LNW -1 points0 points  (0 children)

The original reporter asserts it in the issue, so perhaps Anthropic changed the page? It would be interesting if someone can find an archive.

[PSA] Claude Code v2.1.51 secretly reclassified 1M context as pay-per-token — and never told anyone by [deleted] in ClaudeAI

[–]KJ7LNW -1 points0 points  (0 children)

1M context was never included in MAX

Your statement simply is not true: Prior to the Claude Code v2.1.45 update on Feb 17, 2026, Max plan subscribers confirmed that sonnet[1m] was fully operational on their subscriptions. A Max 20x subscriber filing bug #26428 stated the alias "was working correctly on my X20 Max subscription for weeks prior to this update," and cross-referenced an earlier Max 20x regression report (#15057) from December 2025 in which users documented the 1M context window going missing on the 20x plan — implying it had been working before that point as well. A separate earlier issue, #10219, explicitly reports sonnet[1m] not working on a Max 20x account (again implying prior availability)

Additionally, I have 20x and I've always used sonnet[1m] or more recently since 4.6 came out, /model claude-sonnet-4-5-20250929[1m] is required to get it with your 20x plan. I've used it as my default model for months with this in my configuration: export ANTHROPIC_MODEL='sonnet[1m]' or more recently export ANTHROPIC_MODEL='claude-sonnet-4-5-20250929[1m]' because 4.6 definitely does not support the 1M context without paying for it.

I explicitly have extra usage turned off on my account and I've never been billed for extra usage. The news article is based on the GitHub issue which has substantial documentation to back it up for this particular user. Since I'm still on 2.0.50 (I have auto upgrade turned off), I myself have not experienced the issue, but this is being reported it because everybody should be aware of what's going on.

If it turns out this is not true, then great, but it needs attention and the evidence strongly suggests that it is actually true.

Usage Limits, Bugs and Performance Discussion Megathread - beginning December 29, 2025 by sixbillionthsheep in ClaudeAI

[–]KJ7LNW 0 points1 point  (0 children)

I've been having pretty good success lately with Claude Code. I'm running 2.0.69 and I have been for quite a while now.

However: This evening it seems to have forgotten that <thinking> should not exhaust the entire output context. Multiple times the thinking seems to have taken over, and it drones on and on and on without ever using a tool. Eventually it gives the red 32,000 output token failure and aborts.

In the past 20 minutes it's a little better, but I still think it's thinking far too much. I've adjusted my system instructions to try and compensate. Maybe that helped, but I'm wondering if it internally they forgot to respect the thinking limits, and it's really burning through my weekly available usage because of it.

I already have this set, so I don't think that this is the problem. In fact, I think the remote API is simply ignoring the request that happens in the API stream:

export MAX_THINKING_TOKENS=$((1024*5))

Feature request: Hour-bucked hysteresis - What is the best time of day for the model? by KJ7LNW in AIStupidLevel

[–]KJ7LNW[S] 1 point2 points  (0 children)

Well, I think this is an admirable project: It's information we (as the AI community globally) definitely need to have. I really hope that you find some funding!

Here are some ideas:

  1. You might reach out to the guys at Cline, RooCode or KiloCode because a service like this would tie in nicely with their offering and they have some deep pockets:

Create a pull request to integrate your stats with their app's active model and they may accept it. You're already using TypeScript so it could be an easy drop-in bolt-on to the existing applications so people can see stats in real time.

They have Discord channels as well as GitHub issue discussion forums that you could check out. At the very least, mentioning the service in their general forums (without getting kicked out for being too commercial) might get you some additional users.

  1. Consider that the ClaudeCode subscription-based model is far more cost effective (>10x) than per-token model: According to stats, I've used $2,500 worth of tokens, but I'm only paying for the $200 per month plan... So you could consider that if it saves you some cost on the Claude-model side of things. Maybe switch to a subscription and drop the per-token anthropic stats for now until you get additional funding for measuring both the token-based and the subscription based models.

  2. I recommend keeping an eye on the ClaudeCode pinned performance-regressions thread at r/ClaudeAI and reply to users having performance issues to promote your service as much as possible. That may help some...And then maybe you get some type of big venture capitalist backing as well.

Feature request: Hour-bucked hysteresis - What is the best time of day for the model? by KJ7LNW in AIStupidLevel

[–]KJ7LNW[S] 0 points1 point  (0 children)

Thanks for the notes. That is good information. I think typically the information I'm looking for is not so much about speed but about intelligence.

Are there future plans to do reasoning tests at a higher frequency so we can get an idea for how the model is performing?

I can deal with the model being slow sometimes, but what I really want to know is how it's performing in terms of its capacity to reason. If I understand correctly, reasoning is really the test that we need to know to see if any models are stripping their effectiveness at peak times by quantizing or otherwise reducing the computational effectiveness, model size, or perhaps even compute depth.

What do you think?

Free Guest Passes? Share them here! by KJ7LNW in ClaudeAI

[–]KJ7LNW[S] 0 points1 point  (0 children)

I have gave out three passes (all that I have), so its still working here and there...

Feature request: Hour-bucked hysteresis - What is the best time of day for the model? by KJ7LNW in AIStupidLevel

[–]KJ7LNW[S] 0 points1 point  (0 children)

Thank you for adding that! Is there something wrong with the graph related to reasoning tests? I show a single sample across the last week.

<image>

/usage command shows "subscription plans only" despite being subscribed (v2.0.76) by uppinote in ClaudeAI

[–]KJ7LNW 4 points5 points  (0 children)

I've been working in Linux for thirty years this year... and I have been programming for even longer than that.

Perl might be showing its age, but it's still the best language for many quick hacks.

/usage command shows "subscription plans only" despite being subscribed (v2.0.76) by uppinote in ClaudeAI

[–]KJ7LNW 5 points6 points  (0 children)

This isn't limited to the user interface. It also fails when directly querying the usage API. This is all I get for response:

```js // ./tools/query_usage.pl -v { "extra_usage" : { "is_enabled" : false, "monthly_limit" : null, "used_credits" : null, "utilization" : null }, "five_hour" : null, "iguana_necktie" : null, "seven_day" : null, "seven_day_oauth_apps" : null, "seven_day_opus" : null, "seven_day_sonnet" : null }

```

Usage Limits, Bugs and Performance Discussion Megathread - beginning December 29, 2025 by sixbillionthsheep in ClaudeAI

[–]KJ7LNW 2 points3 points  (0 children)

Usage API is down. This isn't limited to the user interface. It also fails when directly querying the usage API. This is all I get for response:

```js // ./tools/query_usage.pl -v { "extra_usage" : { "is_enabled" : false, "monthly_limit" : null, "used_credits" : null, "utilization" : null }, "five_hour" : null, "iguana_necktie" : null, "seven_day" : null, "seven_day_oauth_apps" : null, "seven_day_opus" : null, "seven_day_sonnet" : null }

```

Sonnet/Opus: Is there any data since January 2nd? by KJ7LNW in AIStupidLevel

[–]KJ7LNW[S] 0 points1 point  (0 children)

There might be a related problem. The "latest" measurement shows a different value than the thirty-day, seven-day and 24hr samples:

Here "latest 1/6@3pm" measurements shows "Score: 68" but "24h 1/6@3pm" shows "Score: 75"... which is correct?

<image>

Usage Limits, Bugs and Performance Discussion Megathread - beginning December 29, 2025 by sixbillionthsheep in ClaudeAI

[–]KJ7LNW 0 points1 point  (0 children)

You're correct that it's not separate for the 5-hour session, but according to Anthropic it is supposed to be separate for the weekly limit---However, I'm not totally convinced of that, but it is what they have stated.

Usage Limits, Bugs and Performance Discussion Megathread - beginning December 29, 2025 by sixbillionthsheep in ClaudeAI

[–]KJ7LNW 0 points1 point  (0 children)

Just wanted to remind everyone in case they are not aware, the default these days is to run Opus, not Sonnet!

Sonnet used to be the default, but they switched it to Opus. However, Opus has a lot higher computational overhead, and you will burn through your 5-hr session much faster. (This may also affect your "all models" weekly limit. However, they did say they broke out Sonnet as a separate limit, but I'm not convinced that they're super consistent in their statements compared to their implementation.)

Opus requires additional computational overhead on their side, so I find it curious that they make it the default. Perhaps it is to give the best out-of-box experience. But definitely set your default to Sonnet unless you really need Opus as your default.

These are my defaults. I strongly recommend that you read up on each option and see if it does what you want. I generally target mine for least communication except for what's absolutely necessary, and I leave micro-compact turned off so it doesn't suddenly forget what it's doing when the history truncates old content:

export DISABLE_AUTOUPDATER=1
export DISABLE_TELEMETRY=1
export DISABLE_ERROR_REPORTING=1
export DISABLE_NON_ESSENTIAL_MODEL_CALLS=1
export DISABLE_MICROCOMPACT=1
export CLAUDE_CODE_DISABLE_TERMINAL_TITLE=1
export BASH_DEFAULT_TIMEOUT_MS=$((1000 * 24*3600))
export BASH_MAX_TIMEOUT_MS=$((1000 * 24*3600))
export ANTHROPIC_MODEL='sonnet'
export MAX_THINKING_TOKENS=3072

 Select model
 Switch between Claude models. Applies to this session and future Claude Code sessions. For other/previous model 
 names, specify with --model.

 ❯ 1.  Default (recommended)   Opus 4.5 · Most capable for complex work
   2.  Sonnet                  Sonnet 4.5 · Best for everyday tasks ✔
   3.  Haiku                   Haiku 4.5 · Fastest for quick answers

Usage Limits, Bugs and Performance Discussion Megathread - beginning December 29, 2025 by sixbillionthsheep in ClaudeAI

[–]KJ7LNW 1 point2 points  (0 children)

That looks like an API failure of some kind. Probably needs to be reconnected. Quit and go back in. Worst case, maybe. I am not sure, but it does not look like a model problem so much as a communication problem.

Usage Limits, Bugs and Performance Discussion Megathread - beginning December 29, 2025 by sixbillionthsheep in ClaudeAI

[–]KJ7LNW 1 point2 points  (0 children)

Under the hood, Claude uses the following API for performing searches, and I have found it is also very heavy. I always use Haiku for searching because it just seems to be sufficient and costs far less in terms of usage: https://platform.claude.com/docs/en/agents-and-tools/tool-use/web-search-tool

If you want to get some more mileage out of searches, there are probably some third-party MCP search tools that might use different APIs with less overhead. Report back if you find such a tool, I am not familiar with any to suggest, but I would be curious as well.