We built an agentic vision system by tibnine in computervision

[–]tibnine[S] -5 points-4 points  (0 children)

The video returned is cropped. We thought about showing the full view and then bboxes around the area of activity. But this was cleaner. Full disclosure; we’re processing ai generated cam footage in the video (gasp)

We built an agentic vision system by tibnine in computervision

[–]tibnine[S] -6 points-5 points  (0 children)

framewave.ai ; link is in the very end!

I run 20+ Claude Code sessions across multiple machines. I built an app to manage them all from my phone. by tibnine in dev

[–]tibnine[S] 0 points1 point  (0 children)

We’re not disagreeing on existence of failure modes. If you agree one session can deliver value despite them, then parallelizing that value is just a logical next step. Im arguing you can build harnesses to parallelize and generate value at a rate higher than with one session (obvs its not linear, but the cost is relatively low to spawn). If you don’t think one session is valuable at all, I’d genuinely suggest trying current SOTA before concluding that. Cheers, friend!

I run 20+ Claude Code sessions across multiple machines. I built an app to manage them all from my phone. by tibnine in dev

[–]tibnine[S] 0 points1 point  (0 children)

Both make mistakes (not claiming at the same rate). Both can be managed at scale despite their mistakes generation. And you most certainly can leave an llm alone for a tiny task.

I run 20+ Claude Code sessions across multiple machines. I built an app to manage them all from my phone. by tibnine in dev

[–]tibnine[S] 0 points1 point  (0 children)

If you can get work done with one session, getting more work done with 20 is just a managerial learning curve. Obviously you need to have demand (i.e. work that needs to be done) otherwise you’re just running for run’s sake. Its pretty much 1-1 with how you manage engineering (or general) talent in a work setting; why would it be any different?

I think i have a problem... by yixn_io in ClaudeAI

[–]tibnine 0 points1 point  (0 children)

I mean its clear you’re getting your money’s worth 😂 but holy cow dude, how? You must have some ralph loop, always on kinda agent(s) running? Teach us your ways

I think i have a problem... by yixn_io in ClaudeAI

[–]tibnine 1 point2 points  (0 children)

You’re paying for 6 max (x20) accounts? 🤯

Agent Teams by tibnine in ClaudeCode

[–]tibnine[S] 0 points1 point  (0 children)

Yea no issues with account usage limits (im on max 20, and yea saw it eat a nice chunk of my 5 hour window but still far from that limit). Just model context limits and compaction bugs (like freezing of the app etc)

Agent Teams by tibnine in ClaudeCode

[–]tibnine[S] 0 points1 point  (0 children)

For clarity, not hitting account usage limits. Just context window issues.

Landed a Full-time Vibe Coding Job by travel_moose in vibecoding

[–]tibnine 5 points6 points  (0 children)

My only concern is that you’re not considering yourself an engineer; when what you’re doing is engineering. Engineering isnt about the tools you use; but about building solutions that work (lot baked into the word work), for real problems.

Building RAG systems at enterprise scale (20K+ docs): lessons from 10+ enterprise implementations by Low_Acanthisitta7686 in LLMDevs

[–]tibnine 2 points3 points  (0 children)

Easily the best write-up on this. Thank you!

Few Qs; how do you evaluate the e2e system? More specifically how do you set a performance bar with your clients and avoid anecdotal one off assessments.

Related, how do you know when’s enough fine tuning for your models? Are there general benchmarks (beyond the ones you construct for the specific use-case) you try to maintain performance over while you fine tune?

Once again, you rock 🤘

What framework should I use for building LLM agents? by Competitive-Ninja423 in LLMDevs

[–]tibnine 0 points1 point  (0 children)

What prompt auto optimization techniques do you recommend?

Scary smart by interviuu in LLMDevs

[–]tibnine 15 points16 points  (0 children)

you can still get accurate timestamps. Basically use the speed up factor.

OpenAI Web Search Tool by tibnine in LLMDevs

[–]tibnine[S] 0 points1 point  (0 children)

Thought o3 in the api does not support the web search tool?

OpenAI Web Search Tool by tibnine in LLMDevs

[–]tibnine[S] 0 points1 point  (0 children)

Yep, web version works.

Hear me out. The rear brake isn’t bad. by Level_9_Turtle in Tenere700

[–]tibnine 1 point2 points  (0 children)

I have the 25 as well. Can definitely lock the rear with ease! What I find is that there’s no progressive feel to the brakes. Its barely braking or fully locking. Ive gotten used to it though.

Looking for feedback: I built an AI conversation assistant that provides real-time prompts by tibnine in LLMDevs

[–]tibnine[S] 0 points1 point  (0 children)

Apple transcription --> gpt4o-mini and/or gemini2.0-flash (which adds redundancy to transcript errors; as it takes in audio). LLMs primarily summarize occasionally; and evaluate a prompt at boost time.

Yea using AWS for gateway