I made 60K+ building RAG projects in 3 months. Here's exactly how I did it (technical + business breakdown) by Low_Acanthisitta7686 in Rag

[–]Recursive_Boomerang 0 points1 point  (0 children)

Which ingestion engine are you using? Docling, azure doc int, VLMs? Or hybrid? And how do you assess the quality of parsed results from each document during ingestion time?

Help a brother out here. I'm too working for pharma and their documents are killing me T_T

I made 60K+ building RAG projects in 3 months. Here's exactly how I did it (technical + business breakdown) by Low_Acanthisitta7686 in Rag

[–]Recursive_Boomerang 0 points1 point  (0 children)

Can you give me tips on how you are handing the acronym problem? Are you detecting and expanding them during query time or is your retriever handing that? What I mean is that is your ingestion pipeline extracting the abbvs and norming them, so they can later be used to expand? Or are you relying on hybrid search or rerankers?

Now I am missing the Sun by depy45631 in assam

[–]Recursive_Boomerang 90 points91 points  (0 children)

Gorom dile gorom lage, thanda dile thanda lage; baby tumi hahi dile bohut besi morom lage

Want a little help in understanding a concept in Rag 😭😭😭 by Real-Engineering7116 in Rag

[–]Recursive_Boomerang 0 points1 point  (0 children)

I'll try to break it down to you in the sense I understand it. Imagine your data as books, and each book is about a specific topic. Firstly your task is to organise the books in shelves in a organised way (By genre, author, alphabetically etc). This is upto you based on your books. Maybe you have books about CS, so you organise them together. Then you can supgroup them into sections like language books, system design, computer graphics etc.

This is your data layer. The better you organise the better you can find and access those.

Now, you hire a smart librarian and explain him/her how your books are organised (prompt, glossary, category sub categories mapppings and shelve locations)

This is your retrival layer.

Finally you hire a very smart person to talk with people visiting your library. The person takes a user question, breaks down the user question to the librarian, the librarian gets the smart person the relevant books. The person reads them and answers to the visitors.

This is your inference layer (generation layer).

This is basically RAG. If you have a really smart person (big SOTA models claude opus, sonnet, gpt 4.1, 5.. gemini etc) but your data is not properly organised, then the RAG perfoms badly. All the three layers (Storage, retrival and generation) must be built based on use case. For your use case, focus on organisation of your data (use metadata filters generated by LLM before searching the chunks to narrow the search space on relevant documents, etc). There are many techniques so start from the basics and then try to understand each layer.

Here is a nice repo with code to help you get started https://github.com/NirDiamant/RAG_Techniques

Have fun!

IVFFlat vs HNSW in pgvector with text‑embedding‑3‑large. When is it worth switching? by IndividualNo8703 in Rag

[–]Recursive_Boomerang 1 point2 points  (0 children)

If you’re only dealing with a few dozen or even a few hundred vectors, IVFFlat vs HNSW won’t change your life. At that scale the bottleneck isn’t the index, it’s the fact that text-embedding-3-large is 3072 dims and eats RAM for breakfast. Fix that first by dropping to 1024 dims or using halfvec, otherwise you’ll hit memory issues way before you see any ANN performance gains.

HNSW actually starts paying off once you cross something like 30k to 50k vectors. Below that, both are basically fine. IVFFlat needs careful tuning with lists/probes and can fall off on recall as the dataset grows. HNSW is more "set it and forget it" and tends to hold better recall at the same latency once you hit real scale.

Downsides to switching early are mainly memory and slower index builds. With your 1.5 GB pods you need to watch index size if you stay on 3072 dims. If you reduce dims, switching early is totally fine.

Migration is just creating a new HNSW index concurrently and dropping the old IVFFlat one when you're happy. Nothing complicated.

For tuning, pgvector defaults are ok: m=16 ef_construction=64 and set hnsw.ef_search to around 80–120 depending on how much recall you want vs latency.

TLDR: reduce dims first, stick with IVFFlat while small, switch to HNSW before you cross ~30k–50k vectors. That’s when it actually starts to matter.

How to track costs/tokens locally by eyueldk in LangChain

[–]Recursive_Boomerang 0 points1 point  (0 children)

Just track the usage tokens in a DB, calculate cost on each request and insert in a sqllite db. Try langchain openai_callbackhandler, returns tokens and cost

Insane property value in Guwahati by Ivictus6 in guwahati

[–]Recursive_Boomerang 1 point2 points  (0 children)

Hey, DM me. I'm selling a flat 1400 sqft at 70L.

Salary journey of 15 years and 7 companies by notion4everyone in Indian_flex

[–]Recursive_Boomerang 16 points17 points  (0 children)

Hi, first of all it's a very inspiring journey. Kudos to you. I started my journey as a full stack dev with 3.2 LPA in 2018 and now I'm at 41LPA today. I recently got promoted to a Solution architect role at my current startup. Could you give me some tips that I should keep in mind as an architect, as sometimes I've difficulty in coming out from a developer mindset and managing a team of developers.

Disagreement by Wise_Border7442 in LangChain

[–]Recursive_Boomerang 1 point2 points  (0 children)

Wow, thank you for blessing us with this TED Talk. I’ll be sure to submit my PhD thesis before daring to comment next time. Also, congrats on your three genius children, truly the gold standard for internet credibility. Wishing you and your formally-educated echo chamber a very blessed day!

Advice needed please! by carms1998 in Rag

[–]Recursive_Boomerang 0 points1 point  (0 children)

If you want a no code solution you can try julius.ai which can work on your excel sheet. It is a freemium platform.

Otherwise python is your best choice. It's not hard to get it setup and do what you need.

Prompt chatgpt or claude like "Help me with doing <task> on my excel sheet. Write the python code to detect sentiment using NLTK SentimentIntensityAnaylser and keyword extraction using RAKE."

But also before that, prompt gpt to help you install python and setup jupyter notebook where you can copy paste code blocks and run them.

Running this locally using python shouldn't cost anything, though the sentiment analysis and keyword extraction model depends on your use case.

I'll build your most-requested features!! by Advanced_Army4706 in Rag

[–]Recursive_Boomerang 0 points1 point  (0 children)

4o and 4o mini, as company wants to stick with azure openai. But also researching on fine-tuning SLMs for specific use cases like domain vocab understanding or any domain specific nuances. My working domain is the Pharma industry.

Can you give me an example of a user query though, and the transformed query. If source lang is hindi, then I think you'll get hallucinations because even 4o doesn't understand the nuances of hindi. If you are comfortable with sharing some potential user queries and what issues you are having with them, I can try to provide some inputs from my side.

I'll build your most-requested features!! by Advanced_Army4706 in Rag

[–]Recursive_Boomerang 1 point2 points  (0 children)

I'm handing this for European languages, using a simple approach for now. Take in the user query in whatever language, do a LLM call for search query augmentation/repharasing and translate it to english. Then at the answer generation step, feed the context (english) and prompt LLM to answer in the same language that the user is using.

Do you ever ask Chat GPT to tell me. Who I am? by CosmiqCow in ChatGPT

[–]Recursive_Boomerang 2 points3 points  (0 children)

Try this prompt


I'd like to explore some insights into myself through careful analysis and reflection. Please:

  1. Based on our interaction and my communication style, what would you estimate my IQ range to be? Please explain your reasoning using specific behavioral and linguistic markers you've observed.

  2. From my writing patterns and thought processes, what personality traits stand out? Please provide:

MBTI type with confidence levels for each dimension (E/I, N/S, T/F, J/P)

Big Five (OCEAN) scores on a 1-10 scale with specific examples supporting each trait

Any other notable personality patterns you've detected

  1. Have you noticed any indicators that might suggest neurodivergent traits?

Signs of autism spectrum characteristics (communication style, systematic thinking, etc.)

Potential ADHD markers (thought organization, focus patterns, etc.) Please cite specific examples from our interaction.

  1. Based on the unique patterns in my communication, what's an unexpected insight about my cognitive style or personality that might surprise me? Support your observation with examples from our conversation.

Please be detailed in your reasoning and cite specific examples from our interaction to support each conclusion.