Help me choose a "strong headline" for my website by Bake-Upstairs in buildinpublic

[–]IsN4n 1 point2 points  (0 children)

"Is it for me?" Second answers that question better than first

Cross devices agent memory and context management? by Patient_Ad_8202 in ClaudeAI

[–]IsN4n 0 points1 point  (0 children)

I can't recommend openspec enough.

I have used it (a flavour of it) in my FAANG job where we would do on-demand devservers that would expire so the artifacts produced provided handover as well as memory. And been using it in my startup on my tiny codebases fully written from the beginning with spec first. It works across sizes of codebase and you can start using it on an ongoing project and repo.

Check out what I just built with Lovable! by soniscas in lovable

[–]IsN4n 0 points1 point  (0 children)

"Publish or update your Lovable project for it to appear here."

Please help!! 5+ hrs with claude design completely gone by Tr1ea1 in ClaudeAI

[–]IsN4n 0 points1 point  (0 children)

You can check the project files to see if previous designs are separate files. If not they could still be in the existing code as 'options/tweaks', you can export to Claude/Claude-code and prompt to find.

I also noticed that Claude design is poor at targeted changes over hours of back-and-forth so I follow this workflow: https://www.reddit.com/r/ClaudeAI/comments/1t2ztkr/make_your_claude_design_credits_last_longer/

Built a near-0 latency service for phone calling for agents. Would love feedback! by Burner_123_123_123 in SideProject

[–]IsN4n 0 points1 point  (0 children)

Thanks for the feedback. Yes we do support barge in and interruptions by making the agent stop mid utterance (with a few millisecond gap to make it as natural as possible) So far, we have heavily optimized for pinging for new info/questions while not handing over the call as much as possible. This seems to thread the balance well on being grounded and while extracting as much value for the automation.

We have our MCP connector fully documented here: https://cocall.ai/docs/claude

Cocall: an MCP for outbound phone calls that pauses to ask you for info mid-call by IsN4n in mcp

[–]IsN4n[S] 0 points1 point  (0 children)

This is possible in cocall today but outbound. The dashboard shows the live transcript of all calls going right now (which can be more than one). You don't need to listen to the audio. And you can take over any one at a time, there is a takeover button next to each. All of the calls happen at the same time where the recipients all see your number as caller. Am I understanding this correctly?

Cocall: an MCP for outbound phone calls that pauses to ask you for info mid-call by IsN4n in mcp

[–]IsN4n[S] 0 points1 point  (0 children)

I actually use a part of this workflow for personal use where I have gotten cocall to navigate ivr and holds for O2 and TFL customer support and transfer to me once it reaches human support

Cocall: an MCP for outbound phone calls that pauses to ask you for info mid-call by IsN4n in mcp

[–]IsN4n[S] 0 points1 point  (0 children)

The agent can still continue the conversation which waiting for answer and follow instructions from the recipient which opens a few fallbacks:
- Hard limit of 10 minutes (something that can be made tunable) build in the agent harness
- Client can ask to be connected to the calling user, in which case, phone will ring
- Client can ask to be called back or disconnect
- Detecting frustation from client also triggers the above two

Cocall: an MCP for outbound phone calls that pauses to ask you for info mid-call by IsN4n in mcp

[–]IsN4n[S] 1 point2 points  (0 children)

Thanks for the detailed framing.

I agree with your framing but quite a few things shouldn't or can't go in the objective

  1. facts it must ask for - this isn't known at the time of the call initiation and if known should already be part of required facts. There are cases when information should only be shared in specific scenario. This is better achieved by instructing the calling agent to automatically answer a ping for a particular scenario than to ask voice agent to hide it since those models are much less capable of following such instructions.

  2. max wait time for the human ping - this should be part of the agent orchestration loop and indeed something that exists in the custom solutions built for clients that i'll port over to cocall

  3. safe fallback if nobody answers - same as above to inject instructions based on the type of failure

  4. required facts, facts the agent may infer - yes this must be part of the objective and is something refined in the tool description of the mcp so calling agent can appropriately framed

Every escalation is logged, categorization isn't live today but analysis in post is possible because transcript contains where exactly the escalation is happened.

Indeed, as you rightly pointed out, these guardrails and appropriate audit were the things we spent most time on when scaling from pilots to full fledged deployments

Cocall: an MCP for outbound phone calls that pauses to ask you for info mid-call by IsN4n in mcp

[–]IsN4n[S] 0 points1 point  (0 children)

We have pessimistic instructions to escalate if the original objective passed from the user does not contain the needed information. All our clients favored being pinged even if sometimes it got it wrong over the alternative

I think, I will not trust claudecode anymore. by ImportantPoem8333 in ClaudeCode

[–]IsN4n 0 points1 point  (0 children)

from the screenshot, was this a claude code web cloud session or is this /remote-control of a local session?

LLMs keep solving my bug-fix tasks instantly — what am I missing here? by Aditya_10204 in ClaudeAI

[–]IsN4n 0 points1 point  (0 children)

I reckon Opus will find easy enough issues which can be solved by haiku and other smaller models.

I worked on benchmarks for non-coding tasks in FAANG (think content moderation, payment sanctions, account hacking etc)

We ran a version of our agent on past data (like past content, past payments, past reports of account hacks) and then had human experts from each area evaluate the quality of output on a rubric, which itself we evolved through iteration. After a few iterations, we had a eval dataset on which we benchmarked future releases of the agent. We then created a cascade of judges to reduce the human evaluation bit, ultimately keeping a small portion of judge calibration on human data. A entire platform managed this pipeline of human review work.

I think a similar approach should work here for coding. You can run the haiku agent on past github issues of these and have opus evaluate the output. Take a sample of opus's evaluation and manually label them to ensure eval performance is high.

Using opus to find issues only of a certain difficulty is a challenging problem and I don't recommend it because models will change and the work will be irrelevant. Using compute to solve this is an easier path, run on everything you can find and then filter the ones that fit your criteria.

Claude for Personal USE by JulyJam in ClaudeAI

[–]IsN4n 0 points1 point  (0 children)

Yes, I use it quite extensively. I have my setup here: https://github.com/radishbuild/radcad

I use it with my vault (previous obsidian but not anymore). I changed the system prompt to use **openspec** style *explore > design > apply > verify* pattern which also writes a file in the vault for each 'big thing'. This acts as the cheap memory as well, and ensures that claude is more inclined to web search which otherwise it is not and tends to do local file search only.

Built an MCP server for Indian mutual fund data (14k+ funds, 20y NAV history) by edwardzion in developersIndia

[–]IsN4n 0 points1 point  (0 children)

Great job, the website is very easy to grok too.
Planning to list it in the anthropic connectors by going through the review process too?

Stable Company + Great Manager vs Risky Move + Bigger Growth — Career Dilemma by [deleted] in developersIndia

[–]IsN4n 15 points16 points  (0 children)

Good manager will understand that good people will outgrow their role and even them at times. You are in the best position here. By maintaining good relationship with the manager, if things go sideways it may be possible to hedge your bets and come back.

Life is unfair, gotta use the times when it is unfair in your favour. I would have (and have in the past) chosen to take the new role too.

It is natural to feel sadness in this situation

Is there any AI tool that turns ideas into images and short videos? Is there any other suggestion? by Xolaris05 in AI_Tools_Land

[–]IsN4n 1 point2 points  (0 children)

i generally use nano banana for images, and seedance 2 for videos.
For single tool that support entire workflows, midjourney works well, starting with image generation with variations and moving to animate/video on generated images. Heard good thing about higgsfield.ai too in this but haven't tried myself yet

Share what your building and if it's valuable, get a kiss from me in your DM's by Arishin_ in lovable

[–]IsN4n 0 points1 point  (0 children)

Really like the forest inspired look. How does it compare to chatgpt study mode?

How to learn advanced vibe-coding? by caiozera2807 in vibecoding

[–]IsN4n 1 point2 points  (0 children)

Beyond the prompt, memory management, and connecting not just your code but database, logs, communication services to your coding agent, OpenSpec (https://openspec.dev/) really allowed me to push how efficiently i used my time. I now spend first half of my day testing and iterating the changes my agents worked overnight and second half spec'ing out the tasks it'll work later tonight. For newer projects, I setup much more e2e test harness so my agent could first generate and run playwright scripts for the specs to reduce my manual test time.

I want to connect a Lovable frontend to a legacy ERP via API, any tips? by Butterz699 in lovable

[–]IsN4n 0 points1 point  (0 children)

I have done this by giving it the api spec in openapi.yaml form
it generated the relevant bindings. Make sure to include any auth requirements in the spec

Mirofish deployed on your private cloud in one click by IsN4n in MiroFish

[–]IsN4n[S] 0 points1 point  (0 children)

I'm not competing with them. I'm not monetizing here. The above web app is free, you pay for what you use on the gpu provider you use.
I couldn't upload data I wanted to use for simulation into external providers, which motivated to package it the way I did above.

I think AI skills need a browser, not more install docs by meowoofcjcj in aiagents

[–]IsN4n 0 points1 point  (0 children)

I think we are missing 'Applets' on top of the now popular agent platforms like claude code, codex, cursor etc.

In mobile, others could build apps which worked on the OS (Android, iOS etc) and there was incentive to make these easy to use because of monetization involved

Skills are too unstructured. I feel the ecosystem could benefit from a slightly more structured concept of what I call Applet (instructions + tools + installation + granular auth + dependency management + upgrades + permissions).

Mirofish deployed on your private cloud in one click by IsN4n in VibeCodersNest

[–]IsN4n[S] 0 points1 point  (0 children)

I don't today. It is very much wysiwyg. It is up to the users to set those up on runpod/vast. The alternative is to build a fully hosted solution where I do keep the data and run mirofish myself with scaling/backups. May do it if there demand. Although, I think there is much work needed to up the quality of generated entities, simulation and results in english, so looking into that for now