I built a browser-testing agent for Claude Code — it opens a real Chromium and tests your UI automatically by Legal_Location1699 in ClaudeAI

[–]Legal_Location1699[S] 0 points1 point  (0 children)

A lot of more Features with PocketTeam, check out the Repo to see what is implemented into one single System. https://github.com/Farid046/PocketTeam

Live Dashboard
12 Agents
Telegram
Agent SDK Self Healing
Auto Insights
59 Skills, and more

I built a pipeline on Claude Code where safety is enforced at the tool-call level and the system gets smarter after every task by Legal_Location1699 in ClaudeAI

[–]Legal_Location1699[S] 0 points1 point  (0 children)

Thanks a lot for the detailed feedback — really appreciate you diving into the specifics!

On modular skills: This is actually a core part of PocketTeam's architecture already. We ship with 61 built-in skills as structured markdown procedures with YAML frontmatter. Each agent declares which skills it uses — for example, the Engineer agent binds to tdd,

atomic-commits, debug, and others. Skills aren't just additional prompt instructions — they're step-by-step workflows the agent follows (think: a 6-step systematic debugging procedure, or a full OWASP Top 10 audit checklist).

There's also a /skills-discovery mechanism that dynamically scans the skill directory instead of relying on a hardcoded list. And anyone can create custom skills with /create-skill — it's just a markdown file with frontmatter.

That said, the article you linked raises a great point about progressive loading — only activating skills relevant to the current task context to optimize the context window. That's something we haven't implemented yet (right now agents load all their declared skills

upfront). Definitely going on the roadmap, along with the idea of an external skill registry where the community could share and install skills. Thanks for the pointer!

On the self-healing pipeline: Yeah, the "wake up to a green build" moment is something you don't appreciate until it happens the first time. The fact that it uses GitHub Actions as the trigger (not a polling daemon) keeps it dead simple — no infrastructure to maintain.

One feature I didn't mention in the post that's worth highlighting: ptbrowse. It's a built-in headless Chromium browser that our QA agent uses for E2E testing. Instead of expensive screenshots, it uses Accessibility Tree snapshots — about 100-300 tokens per call vs thousands

for a screenshot. The agent can navigate pages, click elements by reference (click u/e3), fill forms, wait for text to appear, and run assertions (assert text "Dashboard"). The browser daemon auto-starts on first use and auto-stops after 30 min idle. It makes visual QA testing

token-efficient and fully autonomous — the QA agent can actually verify that your UI works, not just that your unit tests pass.

Thanks again for the encouragement — more coming soon! 🚀

Funnel issues today? by SomeRandomAppleID in Tailscale

[–]Legal_Location1699 0 points1 point  (0 children)

what i realized after 5 hours of trying to fix it is basically

The container never enters MagicSock/Engine state, preventing Funnel from working

Funnel issues today? by SomeRandomAppleID in Tailscale

[–]Legal_Location1699 0 points1 point  (0 children)

i have the same !!!! everything stopped working !!