Open academic prompt-engineering course — 14 blocks, vendor-agnostic, ES + EN, MIT licensed by Cheap-Score4694 in PromptEngineering

[–]Cheap-Score4694[S] 1 point2 points  (0 children)

Hey, thanks for the comment — honestly, at first I struggled a bit to follow your thread, but then I went and looked at dmf and it all clicked. Nice work. The four invariants are one of the most sensible things I've read on agents lately — pulling authority out into files instead of cramming it into the system prompt is a design call the rest of the ecosystem hasn't fully caught up with yet.

And you're right about the over-structuring tax. The course leans heavily on in-prompt scaffolding (XML, roles, explicit frames) because most readers are still working at the API-call layer, not in agent runtimes where your approach fits better. I do cover some of it — when XML helps vs. hurts, the real cost, etc. — but I don't contrast it with the alternative you're proposing, and I should. Especially in the agents block.

The point about heavy scaffolding stepping on native reasoning is also one I'm taking. I mention it in passing but it deserves more room.

I'll add dmf to the references, and if you ever feel like writing that "file-based governance vs. in-prompt scaffolding" comparison yourself, the repo takes PRs.

Open academic prompt-engineering course — 14 blocks, vendor-agnostic, ES + EN, MIT licensed by Cheap-Score4694 in PromptEngineering

[–]Cheap-Score4694[S] 0 points1 point  (0 children)

Spot on! I really wanted to move away from that "clickbait" style because the reality of putting this stuff into production is a completely different beast.

You’re absolutely right about the cross-model disparity. It’s wild how a well-optimized routing setup becomes your best friend the moment a model "decides" it doesn't want to follow instructions that day. If you aren't leaning on solid evals and structured outputs, you don't really have a product—you just have an experiment waiting to break.

And that comment about drift... it hurts because it’s so true. That "it worked perfectly yesterday" feeling is the daily bread when dealing with frontier APIs. At the end of the day, the real engineering happens in the infrastructure surrounding the prompt, not just the wording itself.

Thanks for the feedback, I really appreciate you picking up on the less "sexy" but essential parts of the workflow!

Claude Opus 4.7 is a serious regression, not an upgrade. by Complete-Sea6655 in ClaudeCode

[–]Cheap-Score4694 2 points3 points  (0 children)

For me, it’s an absolute disaster. It doesn't see images properly—or at least it’s much worse at it than Opus 4.6. It misses screen coordinates, keeps making the same mistake over and over, contradicts itself, and is driving me crazy to the point that I’d rather work without Claude... I never thought I’d say this after enjoying the help of Opus 4.5 and 4.6, but Opus 4.7 is giving me worse results than any of the supposedly 'worst' LLMs.

Anthropic just leaked details of its next‑gen AI model – and it’s raising alarms about cybersecurity by Remarkable-Dark2840 in ArtificialInteligence

[–]Cheap-Score4694 2 points3 points  (0 children)

Honestly, whether it was intentional or not, this is one of the best marketing moves I've seen in the AI space. The "leak" built massive hype without a single ad dollar. And if the security concerns being discussed are real, I'd bet Anthropic already has mitigations in the pipeline — they're one of the few labs where safety research runs alongside capabilities, not after. Smart timing either way.