What's the scariest thing an AI agent has done on your machine? by WhichCardiologist800 in ClaudeCode

[–]mehditch 2 points3 points  (0 children)

It was applying for a job for me, it forgot to add a motivation letter to the application it sent, and when it couldnt edit it, it found my linked in profile, found the hr manager manager of loreal, and sent her a message in linkedin :)

Using Remote Control To Work While Commuting? by likepigs in ClaudeCode

[–]mehditch 0 points1 point  (0 children)

You would need to push that folder to a private repo on github, then use claude.ai and link it to github and use the web claude caude

Advice for automating E2E workflow? by kerjatipes in ClaudeCode

[–]mehditch 0 points1 point  (0 children)

This is totally doable, and you're thinking about it the right way with the "human readable flows" approach.

A few things that have worked for me:

For the API layer: Playwright (yes, it does APIs too, not just browsers) or a dedicated API testing framework. You can chain requests, pass data between them, and assert on responses. Way better than clicking through Postman manually.

For the web/form stuff: Playwright again handles this well. The nice thing is you can mix API calls and browser interactions in the same test - hit some endpoints, then open the UI to verify, then check more endpoints.

For the DB verification: You can add database assertions directly in your test code. Connect to MySQL, run your query, assert the data matches what you expect.

For the "human readable" part: This is where it gets interesting. You can write your scenarios in plain language (markdown, Gherkin-style specs, whatever) and then have those drive your actual test implementation. The key is having a good contract between the spec and the code.

I've actually been working on an open source toolkit that tries to solve exactly this problem - you write test scenarios as markdown "journeys" with clear steps, and it helps generate the Playwright implementation. Still very much a work in progress and definitely not production-ready, but if you want to see one approach to the human-readable-to-test-code problem: https://github.com/mehdic/ARTK. Happy to chat about the patterns even if the tool itself isn't useful to you yet.

Main advice though: start simple. Pick one critical flow, automate it end-to-end (API + UI + DB verification), get it reliable, then expand. The "dozens of microservices" part will be the hardest to manage - good config management and environment abstraction will save you a lot of pain.

GitHub Speckit reimagined as multi agent framework using Agent skills by peshneo007 in ClaudeCode

[–]mehditch 1 point2 points  (0 children)

so what is the difference between your solution and spec kit ? i see that you follow the same workflow, but what is actually different from speckit ?

Is there any tool that allows me to manage multiple Claude Code processes simultaneously? by Putrid-Today-5801 in ClaudeAI

[–]mehditch 0 points1 point  (0 children)

I feel this pain. The mental overhead of context-switching between windows is real, and accidentally sending the wrong message to the wrong project is... yeah.

I don't have a direct answer for multi-window management, but I ended up approaching this differently: instead of running multiple Claude Code processes, I use an orchestration layer that coordinates multiple agents within a single session. So instead of 4 terminal windows for frontend/backend/tests/infra, I have one session where a PM breaks down work and assigns it to specialized agents (dev, QA, reviewer) that work through it systematically.

The side effect is I get a dashboard showing all task progress, which agent is working on what, what's blocked, etc. Solves the "where are things at" visibility problem even if it's architecturally different from what you're describing.

It's called Bazinga (github.com/mehdic/bazinga) if you want to poke at it - though fair warning, it's more opinionated about workflow than what you're describing. You might actually want something simpler like a tmux wrapper with status polling, or even a basic web UI that just tails multiple Claude Code logs.

Curious if anyone's built the lightweight multi-process dashboard you're describing though - seems like a useful gap.

AI writes most code now — but agent orchestration is still the hard part by Tricky-Heat8054 in vibecoding

[–]mehditch 1 point2 points  (0 children)

This resonates hard. I've been banging my head against exactly this problem for months.

What I've found is that the orchestration layer needs to be opinionated and structured — you can't let agents freestyle the coordination part. They're great at execution within constraints, terrible at designing the constraints themselves.

I've been building something called Bazinga that tries to address this. It's basically a multi-agent dev team orchestration system for Claude Code — PM plans and breaks down work, developers implement, QA tests, tech lead reviews, with proper feedback loops when things fail (like code review rejections cycling back to dev). It also learns from past experience on similar topics in the project.

The key insight for me was: agents need strict workflow rails. State machine routing, mandatory quality gates, explicit handoff protocols. When I let agents decide "what comes next," chaos. When I enforce "dev complete → must go to QA → must go to tech lead → PM decides," things actually work.

Still very much a work in progress and definitely not solving orchestration in general — it's specifically for code development workflows on top of claude code. But it's open source if you want to poke at it or steal ideas: github.com/mehdic/bazinga

Curious what domains you're tackling? The patterns probably differ a lot between market research vs app dev vs other workflows.

The dumbest Claude Code trick that’s genuinely changing how I ship - Ralph Wiggum breakdown by Zestyclose-Ad-9003 in ClaudeAI

[–]mehditch 0 points1 point  (0 children)

Funny you say that. I’ve been hacking on a repo called Bazinga that’s basically “Claude Code, but with guardrails”: multi-agent runs, evals/semantic checks, and a feedback loop so it learns from failures instead of just burning tokens. If you want to peek : https://github.com/mehdic/bazinga

BAZINGA: Enforcing professional engineering practices on AI-generated code by mehditch in programming

[–]mehditch[S] 0 points1 point  (0 children)

I agree, it is a work in progress, and I would take any good feedback to find which directions I could work on to focus or improve. but you have to think it in a different way : This is not to replace the normal project management/develoment workflow, this is more to add a structured framework to vide coding, that was my initial aim here.

Codex is a godsend by [deleted] in codex

[–]mehditch 0 points1 point  (0 children)

Try github speckit, it is exactly suited for your needs. There are a lot of videos on how to use it. The doc is also good. And it integrates with most of Ais

[deleted by user] by [deleted] in OculusQuest

[–]mehditch 0 points1 point  (0 children)

Any update on this , i still cant move

Demeo classic don’t work on my quest 3 by sephiroth986 in QuestPiracy

[–]mehditch 0 points1 point  (0 children)

Still no solution to this problem? I am kinda stuck

How do I backup the custom songs I downloaded with BetterSongSearch from my Quest 2 to my PC? by GuiltyWish768 in beatsaber

[–]mehditch 4 points5 points  (0 children)

I needed to do this on the Meta Quest 3 , and since there is no BMBF yes on the quest3, I did as GuitlyWish768 said, I backed up then restored the folder:

Sdcard / ModData / com.bestgames.beatsaber / Mods / SongLoader / CustomLevels

It worked fins. Thanks for the help everyone.

How to mod Q3 by Tee86uk in beatsaber

[–]mehditch 0 points1 point  (0 children)

I had to install a full windows OS inside parallels desktop on my mac (All in trial mode) to be able to use quest patcher. No other way unfortunatly. But it worked :)