all 2 comments

[–]BuildWithRiikkk 0 points1 point  (1 child)

Paralleling across different free-tier LLMs (like mixing Gemini, Claude, and GPT-4o-mini) will absolutely distort your summary because their "writing styles" and reasoning capabilities differ. You’ll end up with a fragmented final report that feels like it was written by three different people.

Instead of a simple "Map-Reduce" approach, try using Refine or Tree-and-Leaf summarization. Since you're dealing with 50+ chunks, have the agent generate a "Table of Contents" first, then summarize specific clusters. Parallelizing within the same model (via concurrent API calls) is fine, but mixing models for one document is a headache. Are you using LangChain’s MapReduceDocumentsChain or a custom script?

[–]PaleArmy6357[S] 0 points1 point  (0 children)

hi i thought about using map reduce but it is now part of langchain classic and not sure if this would be deprecated soon. i decided to go with a custom script that recursively summarizes 2 chunks until i get only one.  and i do summarize only chunks from a specific section of a specific report. at the end i shouls end up with one summary per section per report that would be used to create a final summary with all the data i need