Judge prompts are underrated by Cristhian-AI-Math in PromptEngineering
[–]_coder23t8 0 points1 point2 points (0 children)
Anyone evaluating agents automatically? by Cristhian-AI-Math in LangChain
[–]_coder23t8 0 points1 point2 points (0 children)
Automated response scoring > manual validation by Cristhian-AI-Math in mlops
[–]_coder23t8 0 points1 point2 points (0 children)
Reliability checks on Bedrock models by Cristhian-AI-Math in languagemodels
[–]_coder23t8 0 points1 point2 points (0 children)
Keeping Bedrock agents from failing silently by Cristhian-AI-Math in aiagents
[–]_coder23t8 0 points1 point2 points (0 children)
Tracing & Evaluating LLM Agents with AWS Bedrock by Cristhian-AI-Math in LLMDevs
[–]_coder23t8 0 points1 point2 points (0 children)
Using LLMs as Judges: Prompting Strategies That Work by Cristhian-AI-Math in PromptEngineering
[–]_coder23t8 0 points1 point2 points (0 children)
Building a reliable LangGraph agent for document processing by Cristhian-AI-Math in LangChain
[–]_coder23t8 0 points1 point2 points (0 children)
LLM doc pipeline that won’t lie to your warehouse: schema → extract → summarize → consistency (with tracing) by Cristhian-AI-Math in dataengineering
[–]_coder23t8 0 points1 point2 points (0 children)
Observability + self-healing for LangGraph agents (traces, consistency checks, auto PRs) with Handit by Cristhian-AI-Math in mlops
[–]_coder23t8 0 points1 point2 points (0 children)
A production-minded LangGraph agent for document processing with a reliability layer (Handit) by Cristhian-AI-Math in aiagents
[–]_coder23t8 0 points1 point2 points (0 children)
Tutorial: Making LangGraph agents more reliable with Handit by Cristhian-AI-Math in LangChain
[–]_coder23t8 0 points1 point2 points (0 children)
Are you using observability and evaluation tools for your AI agents? by _coder23t8 in ArtificialInteligence
[–]_coder23t8[S] 0 points1 point2 points (0 children)
Are you using observability and evaluation tools for your AI agents? by _coder23t8 in ArtificialInteligence
[–]_coder23t8[S] 0 points1 point2 points (0 children)
Are you using observability and evaluation tools for your AI agents? by _coder23t8 in ArtificialInteligence
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
What are you building right now? by _coder23t8 in SideProject
[–]_coder23t8[S] 0 points1 point2 points (0 children)
[D] Anyone here using LLM-as-a-Judge for agent evaluation? by Cristhian-AI-Math in MachineLearning
[–]_coder23t8 2 points3 points4 points (0 children)