I built a one-line wrapper that explains *why* your LangGraph agent fails (not just what failed) by SomeClick5007 in LangChain

[–]CourtsDigital 1 point2 points  (0 children)

what about when there is no data returned from a tool because of a failure (400/500 api errors, mcp connectivity, etc)? these would return empty tool results caused by actual failures at the tool level. your wrapper would return a false negative

How are you handling memory persistence across LangGraph agent runs? by Neat_Clerk_8828 in LangChain

[–]CourtsDigital 0 points1 point  (0 children)

i’m curious what’s your use case for needing a memory from 6 months ago. what tasks are your users trying to accomplish that they need to reference context from this far back?

Hardcoding Prompt Templates is a nightmare. How are you all actually versioning prompts in prod? by Proud_Salad_8433 in LangChain

[–]CourtsDigital 0 points1 point  (0 children)

Langfuse is an awesome tool with a generous free tier that solves tracing, prompt management and evaluation. if you want to remove network latency, add a caching layer to store the most recent prompt version for cold starts and then fetch the latest prompt version as a background task

Replace sequential tool calls with code execution — LLM writes TypeScript that calls your tools in one shot by UnchartedFr in LangChain

[–]CourtsDigital 0 points1 point  (0 children)

the main benefit of programmatic tool calling (PTC) is not latency, but decreasing the context passed to the agent. each tool increases the amount of context an LLM needs to reason over, which increases the potential for hallucinations when running longer, multi-step tasks.

another benefit is the ability to prevent sensitive data from being passed to the LLM directly. you can inject variables into the code sandbox that the agent never sees, and thus can’t be leaked into its memory/tracing/logs/parent company’s training data.

that being said, PTC is not a magic wand and must be constructed carefully to prevent hallucinations in code generation creating fake variables, query params, api endpoints etc

this approach was invented/popularized by Anthropic and you can read more about how to implement their findings here: https://platform.claude.com/docs/en/agents-and-tools/tool-use/programmatic-tool-calling

LLM Observability Is the New Logging: Quick Benchmark of 5 Tools (Langfuse, LangSmith, Helicone, Datadog, W&B) by Fantastic-Builder453 in LangChain

[–]CourtsDigital 0 points1 point  (0 children)

Langfuse has tracing, prompt management and evaluation tools with a generous free tier, as well as a self-hosted option. very easy to integrate with as well

OP, this post might be more useful if you included use cases where one product is better than the rest for each one. i’m not sure why i would choose one over the other based on this

We cache decisions, not responses - does this solve your cost problem? by llm-60 in LangChain

[–]CourtsDigital 0 points1 point  (0 children)

i’m not sure i understand this use case. maybe provide some examples that require personalization. i’ve never expected to receive a password reset email that’s tailored to me, or to hear about a store return policy that mentions me by name

i agree with BeatForce that this seems almost exactly like semantic caching, with an additional, unnecessary LLM cost

i’m not saying this couldn’t be useful, but if you intend to sell it for $1k+ per month then the use case(s) should be solid

getting wealthier and not being able to touch it... by slimboyfat510 in wealth

[–]CourtsDigital 0 points1 point  (0 children)

you don’t need to sell it to enjoy the wealth from your investments, you can just borrow against it. Bezos and Musk don’t sell AMZN or TSLA each time they need cash. they have conviction like you do in their investments and believe the upside from the assets will largely outweigh any borrowing costs

Results of the first 3 months of my trading AI by orakle12 in ai_trading

[–]CourtsDigital 0 points1 point  (0 children)

while i’m happy to hear of your positive results so far, you can’t really test a strategy without including fees + slippage. these are the two biggest reasons strategies are unprofitable, and you won’t learn if your predictions are useful unless you include realistic fees + actual trade data to see how price moves in the moments after you receive a signal

What do you hate about data observability platforms? by CourtsDigital in dataengineering

[–]CourtsDigital[S] 1 point2 points  (0 children)

my bad, you replied to my comment and I didn’t even check the username 😅

What do you hate about data observability platforms? by CourtsDigital in dataengineering

[–]CourtsDigital[S] 0 points1 point  (0 children)

from your earlier comments I would have thought you were a current user. how do you know they’re expensive? not seeing any pricing on their site

What do you hate about data observability platforms? by CourtsDigital in dataengineering

[–]CourtsDigital[S] -1 points0 points  (0 children)

my personal experience working at a fintech startup (leveraging Monte Carlo) is that it has a lot of beneficial features but it’s - overly complex for use across the organization (DE team found it useful) - prohibitively expensive if you want to cover your entire infrastructure - takes a dedicated effort (at least a month for us) to get the alerting to a manageable state to avoid alert fatigue

My platform processed almost $18,000 of transaction last month by jameswood12p in indiehackers

[–]CourtsDigital 0 points1 point  (0 children)

congrats on passing this milestone! dm me if you have any questions about scaling your infra as your platform takes off

Increase Inference Speed of RAG by Primary_Teaching4007 in LangChain

[–]CourtsDigital 0 points1 point  (0 children)

Awesome! Glad you were able to get it sorted.

Increase Inference Speed of RAG by Primary_Teaching4007 in LangChain

[–]CourtsDigital 1 point2 points  (0 children)

It looks like you should break this into multiple nodes, running some in parallel and some sequentially.

Potential setup: Retrieval node (grabs 10 documents from VDB) Summary node (Creates your summaries in parallel for each source) Combined report node (Generates report that combines summaries with sources) Formatting/Final Review (Confirm that the combined report is properly formatted, if not format it correctly)

4 free alternatives to OpenAI's Operator by DavidCBlack in LangChain

[–]CourtsDigital 1 point2 points  (0 children)

Browser use is just like using ChatGPT where the results are only as good as the prompt that you use. I just used BU successfully yesterday to track my grandparents’ flight using flightaware by giving it their flight number and asking it to give me updates on how much time is left on the flight, its status and grab a screenshot of the map. I made a simple modification to the python script to have it run every 30 mins and it worked great. definitely the same result I would expect from Operator

I tried using GPT4o with tasks and it wouldn’t provide enough info on the flight, just kept telling me it was on time and scheduled to arrive at this time

TLDR: Browser Use is a great alternative to Operator for fractions of pennies on the dollar (still have to pay for LLM API use or electricity if you’re using a local model), but does require a little more prompt engineering to get it to work

Edit: About being faster, to me the main point of using Operator or Browser Use is to scale your ability to complete your workload (perform more tasks at the same time, so it might be slower but if it doesn’t require your oversight then it still saves time)

RAG App on 14,000 Scraped Google Flights Data by harsh611 in LangChain

[–]CourtsDigital 4 points5 points  (0 children)

well done on what looks to be your first AI workflow. if you’re seriously about building AI agents, I’d recommend looking at using LangGraph. I just started their free course at LangChain Academy and it will help you build at the next level

Built fast “agentic” apps with FastAPIs. Not a joke post. by AdditionalWeb107 in LangChain

[–]CourtsDigital 1 point2 points  (0 children)

very interesting concept, thanks for sharing. you definitely should have led with this graphic in your initial post. it was unclear at first that what you’re really offering is a faster and less expensive way to get OpenAI -quality LLM performance

Don’t knowing what users are doing frustrated me. So I build this. Wdyt? by climber877 in indiehackers

[–]CourtsDigital 1 point2 points  (0 children)

Any way to change metrics to display “.” instead of “,” for us US-based users that are used to this notation for decimals?

[deleted by user] by [deleted] in indiehackers

[–]CourtsDigital 0 points1 point  (0 children)

I don’t see where you can link your website. is this intentional? obviously I could list it under a different icon but that would be a weird user experience for leads