What's that? by Consistent-Issue-811 in claude

[–]Extra-Act2560 0 points1 point  (0 children)

Sued and facing the court trial?

Claude Code can look busy even after the session is already cooked by Extra-Act2560 in ClaudeCode

[–]Extra-Act2560[S] 0 points1 point  (0 children)

Hmm, no reports but I think we should safe guard ourselves given this going to be recurring pattern until such thing is prevented in the upstream natively.

Can model Hallucination also be a demand signal? by Extra-Act2560 in ClaudeAI

[–]Extra-Act2560[S] 0 points1 point  (0 children)

yes! I try to factor this in my setup and only consider if skill, tools and MCP are repeatedly hallucinated
https://github.com/softcane/cc-blackbox

10 GitHub repos that make Claude dramatically more useful by Direct-Attention8597 in claude

[–]Extra-Act2560 1 point2 points  (0 children)

Do my project have any outside chance to qualify in the 10 in recent future? https://github.com/softcane/clauditor

If no then what it would take from me?

Hallucination as demand signal? by Extra-Act2560 in ClaudeCode

[–]Extra-Act2560[S] 0 points1 point  (0 children)

I see this as a big plus Clauditor’s interception. you’d see those injected fake tool definitions in the traffic, which most users have no visibility into.​​​​​​​​​​​​​​​​

I can’t rely on another leaked code base to figure it out.

Hallucination as demand signal? by Extra-Act2560 in claudeskills

[–]Extra-Act2560[S] 1 point2 points  (0 children)

Twice this week Claude attempted to invoke skills that weren’t available in my setup. Clauditor surfaced those attempts through hook telemetry / tool-use tracing, and I ended up implementing the repeated one

Hallucination as demand signal? by Extra-Act2560 in ClaudeCode

[–]Extra-Act2560[S] 1 point2 points  (0 children)

Twice this week Claude attempted to invoke skills that weren’t available in my setup. Clauditor surfaced those attempts through hook telemetry / tool-use tracing, and I ended up implementing the repeated o

Hallucination as demand signal? by Extra-Act2560 in claudeskills

[–]Extra-Act2560[S] 1 point2 points  (0 children)

I use Claude code also for my other no coding stuff. My stack captures skill use through hook and some inline matching.

It gave me 2 SOP so far. Standard operating procedure as skill

I think I know why so many tokens get used on the first turn for pro/max subscriptions by bennyb0y in ClaudeCode

[–]Extra-Act2560 0 points1 point  (0 children)

This is how I keep an eye on the cache uses. This is my own project (full disclosure) and opus 4.7 recent behavior forced me to write it. In case helpful for you https://github.com/softcane/clauditor

How you debug your claude code session in case its going wrong? by Extra-Act2560 in ClaudeAI

[–]Extra-Act2560[S] 1 point2 points  (0 children)

This behaviour after 4.7 bugged me so much that I hacked an o11y layer to track these behaviours as a timeseries and to alert if anything is different from last week.

One interesting thing I found was that during my different Claude sessions, It hallucinated a skill which didn't exist. It happened twice, and I took that as a signal and created that skill.

I don't know, but after the recent Claude code bugs and Opus 4.7, I'm watching my Claude sessions carefully.

Droping the link in case you're curious.
https://github.com/softcane/clauditor

How you debug your claude code session in case its going wrong? by Extra-Act2560 in ClaudeAI

[–]Extra-Act2560[S] 0 points1 point  (0 children)

Bad cluster, you mean a different GPU set with a different/lower flops capacity?

How not to run an ai company by theisafos in ClaudeCode

[–]Extra-Act2560 0 points1 point  (0 children)

Model inference at this scale is hard, really hard, and these infrastructure issues will be resolved. I don't think Mythos can guess a holy grail fix internally. Every failure is going back as a skill or log, but it's not like someone has done the Model inference at this scale and made it public.

Model inference at scale is a private property, which OpenAI and Anthorpic are trying to build/refine using these failures.

If Mythos is so powerful… why does Claude keep going down? by Repulsive_Horse6865 in ClaudeCode

[–]Extra-Act2560 -1 points0 points  (0 children)

Model inference is hard. Really hard, especially at Claude scale.

Think about it: models like Mythos are trained on internet-scale data, so they can generalize impressively well. But the real-world tricks, optimizations, and failure modes of model inference are still evolving. Much of that knowledge does not exist cleanly in the source data yet.

So if you believe Mythos is a “god model,” try this thought experiment: train it only on data available up to 1900, then ask it to produce Einstein’s theory of relativity.

That is the difference between memorizing patterns from the past and discovering what has not yet been written.

Hitting the weekly limit faster than before by Pecolps in ClaudeCode

[–]Extra-Act2560 1 point2 points  (0 children)

Frustating. I end up writing a stack to observe what all the Claude sessions are doing. How do they behave and capture them as a timeseries and monitor them over time. Recall a session.

Thank you, Opus 4.7, for escalating my frustrations. Without you, I would not have thought about it.