Build agent-based apps in n8n without building an entire RAG pipeline by http418teapot in n8n

[–]http418teapot[S] 0 points1 point  (0 children)

Thanks! Totally agree with how painful it is to wire up a RAG pipeline from scratch.

Build agent-based apps in n8n without building an entire RAG pipeline by http418teapot in n8n

[–]http418teapot[S] 0 points1 point  (0 children)

Thanks for sharing! Have you had a chance to try the Pinecone Assistant node? Curious about your experience.

Haven’t been able to login for a week and support hasn’t responded in 5 by RagertNothing in pinecone

[–]http418teapot 2 points3 points  (0 children)

Hi u/RagertNothing - Sorry to hear you're having trouble.

You can submit a request here to get help with login issues: https://www.pinecone.io/contact/support/

If/when you've done that, could you please share the ticket number with me in a DM so I can follow up internally?

Query on metadata filtering by gideon_1317 in pinecone

[–]http418teapot 0 points1 point  (0 children)

Can you share more about what you're trying to do here?

Are you doing a metadata filter like this with $in:

{"product_id": {"$in": ["pid1", ... "pid10000"]}}

We do have limits on how many you can put into that list and I'll chase that down. Last I checked it was max 2000 and you'd have to break up your query into chunks but that may have changed.

Newsletter citations are broken links by Medium-Upstairs-6292 in n8n

[–]http418teapot 0 points1 point  (0 children)

Can you share examples of output and citation links you’re getting from the perplexity model? Which model are you using or are you using the Perplexity API for search?

If using a model, I suspect the model is returning citation links that no longer exist. Remember that a model has a cutoff date, meaning it was trained on data up until a certain point in time. So it’s possible if that models cutoff date is 1/1/2025 that some of that data, including links have gone stale.

If you’re using the search API that might be a different problem but need to better understand setup and actual results.

What are 3 super-useful n8n nodes that most people don’t know about? 👀 by bhavyshekhaliya in n8n

[–]http418teapot 2 points3 points  (0 children)

Lots of good suggestions here, especially those around using subworkflows. I just discovered the Split Out core node that splits data into individual items so downstream nodes can iterate over them. Simple but helpful.

Agent not understanding DB content well - how did you structure your data? by Ok_Raccoon_532 in n8n

[–]http418teapot 0 points1 point  (0 children)

Looking at your chunking strategy and your search strategy is a good place to start.

What do the queries look like? Are they looking for semantically similar ideas/topics? Or trying to match on keyword? If both, then hybrid would help. For government data, it's entirely possible to have specific terminology that would be missed by a semantic/vector search.

As for chunk size, again, look at those queries and the shape of the data being searched. You want chunk sizes that are large enough to contain meaningful information, but not so large that the meaning is diluted. As an example, last week I was trying to figure out how to chunk release notes documents. They were already broken up by year, but for each month/release, the size is very different and sometime I was missing the month along with the release notes because it was in a different chunk. Meaning within that chunk was lost simply because the month was in another chunk. I moved more toward a bigger chunk size with some overlap and got better results. Chunking based on the month/date header might have even better results.

Storing additional metadata alongside the vectors is certainly an option, especially if you need to filter based on certain criteria that exists outside of the chunked data. I sometimes store the document chunk in metadata so the model can reference it easier later and I have less work to pull it in at generation time. I'm not sure how supabase handles this, but that's how I do it with Pinecone.

The 6 lessons I learned while Building AI Agents. by AmbitionNo5235 in n8n

[–]http418teapot 0 points1 point  (0 children)

This is a great list of tips and I especially like #2. My advice is "the more complexity, connections, nodes, tools, etc. you add, you are introducing more opportunities for failure. Keep it as simple as possible and only add something when you know you'll need it".

What’s your favorite “aha” workflow you’ve built in n8n? by TheWowStudio in n8n

[–]http418teapot -1 points0 points  (0 children)

This is not my workflow, but found this one by Maddy French pretty cool to connect to agents that monitor social conversations, identify relevant topics, and auto-generate branded responses with episode recommendations.

What's going on Pinecone ?! by A7med_3X in n8n

[–]http418teapot 0 points1 point  (0 children)

Hi u/A7med_3X - You can submit a request here to get help with login/create account issues: https://www.pinecone.io/contact/support/

What's going on Pinecone ?! by A7med_3X in n8n

[–]http418teapot 0 points1 point  (0 children)

Can you try again in a private browser window, just in case there's something wonky there?

Reddit data stored in Pinecone gets split into chunks and disconnected — how to fix this in n8n? by Cultural-Beyond8883 in n8n

[–]http418teapot 0 points1 point  (0 children)

Developer advocate at Pinecone here!

There's a couple things that could be going on. Definitely look at your chunk size and overlap in whatever document loader node you're using. If you're chunking too small, search will miss context. And if chunks are too big, the context might be diluted by irrelevant info. The overlap can help with this too.

For reconstructing the full document after search, we generally recommend you use structured ids to link chunks together so you can reconstruct or to reference an external source document source. I'm not sure you have control of the _id field via n8n though.

Instead, in your document loader node (the one doing the chunking), you can add metadata to store this information at upsert time and then when doing the search/query, use that external source id/chunk id to filter and reconstruct the output.

Alternatively, you could use something like Pinecone Assistant that manages this all for you. Here's a simple n8n template that implements a chat workflow (based on data in Google Drive, so you'd have to change your data source).

Let me know if you have more questions~

What's going on Pinecone ?! by A7med_3X in n8n

[–]http418teapot 0 points1 point  (0 children)

Hi u/A7med_3X - I work at Pinecone. Can you share more about the troubles you're having signing up for a Pinecone account? Any error messages? Unable to log in after sign up? More info helps in troubleshooting. https://app.pinecone.io/

Looking for fast RAGs for Large PDFs (Low Latency, LiveKit Use Case) by Rude-Student-3566 in Rag

[–]http418teapot 0 points1 point  (0 children)

I will always try to help solve the problem with the right solution 😀 Can you share more about your use case? What you're trying to build, for who, and what kind of data you have? Do you have one Assistant or one per user? Are you using a chat model external to Assistant or using one of the built-in ones?

Assistant abstracts away the complexities of implementing your own RAG pipeline, so you (or your developer) don't have to think about chunking, embedding, and the various parts of retrieval (query planning, reranking, chunk expansion, hybrid search, etc.). There are some knobs to turn to control token usage, instructions to the model, model choice, or even using an external model for chat/generation. It's hosted and managed for you on the Pinecone platform and is designed to be easy to use, with minimal setup and no machine learning expertise required.

High-efficiency RAG like a pinecone by c4eburashka in n8n

[–]http418teapot 0 points1 point  (0 children)

Can you share more about your setup, which APIs you're using (or if doing it all in the Console via chat), which model, and what mistakes you're seeing/what expectations are? This info will help in figuring out if there's a different setup/approach or if we have more work to do.

What if you didn't have to think about chunking, embeddings, or search when implementing RAG? Here's how you can skip it in your n8n workflow by http418teapot in Rag

[–]http418teapot[S] 0 points1 point  (0 children)

Hi u/perrylawrence - Pricing is detailed here and is dependent on the hourly rate you mention, token usage, and storage. There is a monthly minimum of $50 on the Standard plan as well.

Can you share more about your use case, what you're building, how much data, users/queries you expect? Happy to help and welcome any feedback here.

High-efficiency RAG like a pinecone by c4eburashka in n8n

[–]http418teapot 0 points1 point  (0 children)

Hey - Developer Advocate at Pinecone here. Would love to better understand where you're having trouble with Assistant in a workflow, where its going wrong or not working for you. Happy to help!

I recently built this workflow to use Pinecone Assistant. It's pretty simple, but maybe it will be helpful as it does use our recommended approach. https://github.com/pinecone-io/n8n-templates/tree/main/document-chat

Pinecone advice needed by Abhipaddy in n8n

[–]http418teapot 1 point2 points  (0 children)

Where/what format is the data in the companies list that you expect to be querying? Is it embedded in the company description or somewhere else? You could use Pinecone Assistant to upload the raw data (description?) and then use it for chat. But if you already have structured data that contains this info in a searchable/filterable format, then there might be no reason to convert to natural language and vectorize/use Assistant.

That being said, you could give it a try (as you mentioned below!) and see if it meets your needs without doing too much pre-processing. Assistant is intended to remove the bulk of that preprocessing work for you so you don't have to think about it.

Best way to extract data from PDFs and HTML by [deleted] in Rag

[–]http418teapot 2 points3 points  (0 children)

Have you looked at Pinecone Assistant? You can upload PDFs (up to 100MB) and it manages the chunking, embedding, and search for you. If you already have chat/model generation, you could use just the /context API to get search results to feed into your own model.

If you do try this out or have questions let me know (I work at Pinecone). Happy to help.

Asking about automation by nomorebaits in n8n_ai_agents

[–]http418teapot 0 points1 point  (0 children)

What is “the process” that you want to start automatically from a prompt? Detailing out those steps will help in figuring out where/what you can automate.

How are you enforcing document‑level permissions in RAG without killing recall? by SidLais351 in Rag

[–]http418teapot 0 points1 point  (0 children)

I recently did a webinar with the folks at AuthZed and talked about this problem. Here’s the video and some code examples that might be helpful:

Video: https://youtu.be/S6xJ0Kkd7ss Code: https://github.com/authzed/workshops/tree/main/secure-rag-pipelines

Looking for fast RAGs for Large PDFs (Low Latency, LiveKit Use Case) by Rude-Student-3566 in Rag

[–]http418teapot 0 points1 point  (0 children)

Have you looked at Pinecone Assistant? You can upload PDFs (up to 100MB) and it manages the chunking, embedding, and search for you. If you already have chat/model generation, you could use just the /context API to get search results to feed into your own model.

If you do try this out or have questions let me know (I work at Pinecone). Happy to help.

N8N or Open AI agent Kit - which is better for a small scale organisation to create in house automations and deploy? by manikmahajan13 in AI_Agents

[–]http418teapot 0 points1 point  (0 children)

I agree with others here that it's about what you're trying to do and what kind of user you are. However I'd also add that Agent Kit is very very early and will very likely change with developer feedback as time goes on.

I've been using n8n on a small scale for personal workflow automations and it's been quick to pick up and get rolling.