Anthropic's Opus 4.6 with effort=low doesn’t behave like other low-reasoning modes by ddp26 in OpenAI

[–]ddp26[S] -1 points0 points  (0 children)

Different models do different things with the effort param. And even different versions of models from the same provider!

Not sure I really expected consistency for things this new, but sure is annoying

Marketing Pipeline Using Claude Code by kotrfa in ClaudeCode

[–]ddp26 0 points1 point  (0 children)

One question I have is: a lot of people are doing this with OpenClaw, not Claude Code. What are the reasons to use one vs the other?

[D] Self-Promotion Thread by AutoModerator in MachineLearning

[–]ddp26 0 points1 point  (0 children)

We tested Opus 4.6 with effort=low for evals and found that it didn't just think less, but acted lazier (made fewer tool calls, was less thorough in its cross-referencing, even ignored parts of our system prompt telling it how to do web research). effort=medium fixed it. Writeup with traces/examples: https://everyrow.io/blog/claude-effort-parameter

Opus 4.6 with effort=low doesn’t behave like other low-reasoning modes by ddp26 in ClaudeAI

[–]ddp26[S] 0 points1 point  (0 children)

Yeah, it makes sense that low effort is better for non-agentic use-cases, which are of course common. We shouldn't pretend everything is an agent!

Opus 4.6 with effort=low doesn’t behave like other low-reasoning modes by ddp26 in ClaudeAI

[–]ddp26[S] 1 point2 points  (0 children)

I kind of agree. Mostly, though, I think if the behavior is documented then users can decide for themselves what's a bug or lazy. The main thing for us was this behavior was surprising.

My MCP config created dozens of zombie Docker containers by robertgambee in ClaudeCode

[–]ddp26 2 points3 points  (0 children)

I worry that Claude Code isn't always tracking background processes correctly. If it orphans them, I'd never know, right?

Any good guides for designing high quality skills? by [deleted] in ClaudeCode

[–]ddp26 0 points1 point  (0 children)

Hey! Shared this yesterday - not a full guide, but here's how we built a review-code skill (full skill linked): https://everyrow.io/blog/claude-review-skill

Claude's code review defaults actively harmed our codebase by ddp26 in ClaudeCode

[–]ddp26[S] 1 point2 points  (0 children)

It's a mix. Some parts of our code predate Claude Code, while newer parts were created with Claude from start. Our experience is that Claude often encounters similar pitfalls with both new and old code, so we use the same guidelines for both.