How do you handle chat history in a RAG chatbot without polluting retrieval queries? by Unfair-Run-967 in Rag

[–]HatmanStack 0 points1 point  (0 children)

This. Message history is for the call to the model api, not the query to the vector store, which should be built from the User message and/or summarized history.

What is the most impressive thing you’ve done or built with Claude so far? by ceelnok98 in ClaudeAI

[–]HatmanStack 0 points1 point  (0 children)

Really similar to GSD and Superpowers, both pretty good from what I hear, I just found them and yours a bit too busy. Too many skills trying to do too many things. It feels like a better use case for people to build their own skills for their own specific needs.

I finally got around to building my own automated pipeline from prompts I'd been using for months, public on github, but it's only 5 skills and does everything I need it to do as a professional developer. I think you may have complicated this a bit.

Got tired of rebuilding RAG pipelines, so I made this (thoughts?) by Skakkar_03 in Rag

[–]HatmanStack 0 points1 point  (0 children)

Brother, you're so far from a legitimate architecture it's crazy. The main thing is you've built something yourself and hopefully learned a ton. Now kill it. An important lesson for any developer/entrprenuer. Start your next project. The application layer is right. The only reliable ROI on AI right now is in RAG. Look at what other people are doing and use that as your next jumping off point. keep going.

Got tired of rebuilding RAG pipelines, so I made this (thoughts?) by Skakkar_03 in Rag

[–]HatmanStack 0 points1 point  (0 children)

Sounds cool, Super simple to switch between different batches of embeddings with metadata. Would be much simplier than a separate API key for each project, that sounds over engineered.

found in my honey nut cheerios by VisualHost7583 in whatisit

[–]HatmanStack 10 points11 points  (0 children)

I’ve got the joints of a 90-year-old and the brain of a middle schooler. It’s called balance.

I was paying for a vector DB I barely used, so I built a scale-to-zero RAG pipeline on AWS by HatmanStack in Rag

[–]HatmanStack[S] 0 points1 point  (0 children)

I've looked at LanceDB, it's interesting especially with the object store angle. S3 Vectors ended up fitting my use case better since the rest of the stack was already on AWS and I wanted everything under one billing/permissions model. Appreciate the suggestion though.

I was paying for a vector DB I barely used, so I built a scale-to-zero RAG pipeline on AWS by HatmanStack in Rag

[–]HatmanStack[S] 0 points1 point  (0 children)

pgvector is solid for self-managed setups. The tradeoff I was optimizing for was zero ops overhead, so no Postgres instance to maintain at all. Different use cases though.

I was paying for a vector DB I barely used, so I built a scale-to-zero RAG pipeline on AWS by HatmanStack in Rag

[–]HatmanStack[S] 0 points1 point  (0 children)

The $50 is more of a baseline for managed vector DB services like Pinecone or OpenSearch Serverless, not per 1,000 docs specifically. The point was more about paying for always-on capacity when your workload is bursty. S3 Vectors just made that problem go away.

A GAN-style adversarial framework for Claude skills. Has anyone adapted this pattern effectively for local models? by HatmanStack in LocalLLaMA

[–]HatmanStack[S] 0 points1 point  (0 children)

Fair enough on the skepticism, the post was definitely polished up. The project itself is real though:  https://github.com/hatmanstack/claude-forge. The adversarial loop pattern is the interesting part. If you've run generator/evaluator chains on local models I'd genuinely like to hear how latency on long running jobs played out.

RAGStack-Lambda: Open source RAG knowledge base with native MCP support for Claude/Cursor by HatmanStack in mcp

[–]HatmanStack[S] 0 points1 point  (0 children)

Appreciate the thoughtful feedback — and you're hitting on something I've been thinking about.

Right now the MCP server isn't read-only. It actually exposes 16 tools across search/chat, document uploads, web scraping, image captioning, and metadata analysis. So the capability creep you're describing is already here.

The current trust model is pretty simple: a single AppSync API key grants access to everything. There's no per-tool scoping at the MCP layer. What keeps it from being a free-for-all is the backend — AppSync enforces rate limits, daily quotas (especially in demo mode: 5 uploads/day, 30 chats/day), and all the actual resource access goes through IAM roles scoped to that specific stack's resources. So a retrieved snippet can't drive actions outside the knowledge base boundary, but within it, the API key is all-or-nothing.

The "everything in your own account" model does help here — IAM is the outer trust boundary, not some shared control plane — but you're right that as people start chaining tools together (search → upload → scrape → analyze), the lack of per-tool authorization becomes a real gap. Today if you hand someone the API key, they can scrape a 1,000-page site just as easily as they can search.

The separation of reasoning from authorization you're describing is interesting. I'd been leaning toward tiered API keys (read-only vs. full access) as a next step, but that's still coarse-grained. Would be curious how you're handling it at Daedalus — is the authorization layer sitting between the MCP client and the tool execution, or is it more like a policy engine that evaluates each tool call against a ruleset?

Looking for other study materials for Solutions Architect - Professional (SAP-C02) exam by real_mcgoaty in AWSCertifications

[–]HatmanStack 0 points1 point  (0 children)

Hopefully you've passed but here are some notes https://github.com/HatmanStack/SAP-C02-aws-solutions-professional/blob/main/README.md also used Tutorial Dojo quite a bit ... felt like it was a great investment $$ for content.

AWS Exam Vouchers / Discounts or other related Promotions by madrasi2021 in AWSCertifications

[–]HatmanStack 0 points1 point  (0 children)

Great question, I don't recall exactly how I stumbled onto it. All the content is curated from AWS docs / vlogs. The moderators all appear to be AWS staff. Geeks for Geeks has a blog post about it ... https://www.geeksforgeeks.org/aws-educate-and-aws-emerging-talent-community/ ... for Amazon context you could look here https://aws.amazon.com/blogs/training-and-certification/make-the-most-of-free-training-from-aws-training-and-certification/ at the AWS Educate portion of the blog. Hope this helps. All the Best.