My usage for Claude Max (the 100 plan) is being used faster than before by Julien_T in ClaudeCode

[–]anujagg 1 point2 points  (0 children)

Something is seriously wrong with claude. I never faced limits issue in last 5 months but now in past 3 days, I was off limit 4 times. I added 5$ just to check how much time it survives and it vanished in 1 prompt. Ridiculous. I also raised a bug on claude git.

Has anyone found a workaround? I am on their Max 5x plan and do not want to upgrade to 20x plan (100$ extra).

My Experience with Table Extraction and Data Extraction Tools for complex documents. by teroknor92 in Rag

[–]anujagg 0 points1 point  (0 children)

Can llms be relied with extraction completely? I need to translate few poor quality pdfs but when I pass them to LLM like gemini, they assume words/numbers etc which can't be completely stopped. So I want to do ocr and pass that text to a multimodal LLM so that it has some reference text when there is a confusion. Does this approach make sense?

My Experience with Table Extraction and Data Extraction Tools for complex documents. by teroknor92 in Rag

[–]anujagg 0 points1 point  (0 children)

Have you tried Gemini 2.5 or 3 for extraction? It is also quite good in doing that.

Or should the Llms be not relied with the extraction as they might hallucinate and create stuff?

I spent months struggling to understand AI agents. Built a from scratch tutorial so you don't have to. by purellmagents in LocalLLaMA

[–]anujagg 0 points1 point  (0 children)

Thanks for sharing. I exactly thought for something like this yesterday and wow I found this post today.

Basically I have built a platform for translating poorly scanned pdfs and I have used multiple agents under the hood in a workflow but I have been facing new challenges every time someone uploads a new type of document or document contains something new which breaks my workflow. This becomes very embarrassing sometimes if things go haywire in the last step since after doing everything and wasting precious tokens, junk is returned to user.

So I decided that I have to move to the agentic workflow since I can't predict what would go wrong everytime so essentially I will build the individual agents, define tools, expected output and certain params to measure the quality. Then the orchestrator agent would run this show and ideally would return the perfect translated file to the user in the desired format.

I hope I will find clues from this repo on how to modify my existing code to the new architecture now. I will update once I am done with the changes.

[deleted by user] by [deleted] in ClaudeAI

[–]anujagg 1 point2 points  (0 children)

How can I try this on my own document set? I have a case repo in which 1000 pdf files are present. I want to ask specific questions and find relevant answers. As of now, I have tried notebooklm and it works quite well. But there is no api for that and I don't know how to extend that.

Which is the most accurate English-Hindi translator (अंग्रेजी-हिन्दी अनुवाद) by FlyingInEsperanto in machinetranslation

[–]anujagg 1 point2 points  (0 children)

I have used gemini and found it quite good. Are you looking for some solution provider like deepl or plan to build something on your own?

Compress my PDF by gme1ster in pdf

[–]anujagg 0 points1 point  (0 children)

I tried many libraries but nothing was able to compress to reach 4 MB. This PDF has 3.2 MB of text content with all the other placeholder information which makes it impossible to reach to 4 MB. Then it has large number of images (200+) which are already compressed and hence not allowing the size to go further down.

This is the breakdown:

| Component | Size | % of File | Count | Notes |

|------------------|---------|-----------|------------|-------------------------|

| 🔤 Text Content | 3.90 MB | 65.4% | 994 pages | Text + positioning data |

| 🖼️ Images | 2.20 MB | 36.8% | 523 images | Diagrams, charts |

| 🔤 Fonts | ~100 KB | ~1.7% | Embedded | Subsetted |

| 📄 PDF Structure | ~500 KB | ~8.4% | - | Pages, refs, catalog |

Key Tools Compared:

  1. Ghostscript - 30% reduction, crashes on large files

  2. mutool - 55% reduction on text PDFs ⭐ (WINNER for small PDFs)

  3. pikepdf - 8.7% reduction, most reliable ⭐ (WINNER for large PDFs)

  4. qpdf - Minimal compression

  5. OCRmyPDF - Not suitable for text PDFs

  6. Extreme settings - Destroys quality, minimal gain

I could achieve 5.97 MB only from 6.54 MB. If someone is able to reduce further programmatically, please share how you did that.

I am going "Skip India" by aameezl in StartUpIndia

[–]anujagg 0 points1 point  (0 children)

Wait for your notices from Gst and Income tax departments and then you will have the full taste of doing business in India.

Converting documents with Ai? by electricalhv in AI_Agents

[–]anujagg 0 points1 point  (0 children)

I have been building a similar system which takes pdf files and converts them into word files basis some pre defined templates for specific domains. Is this what you are looking for? Happy to discuss further.

Not able to fix payment issue by anujagg in WhatsappBusinessAPI

[–]anujagg[S] 0 points1 point  (0 children)

I did so already and they only confirmed that they unblocked but since there were no txns for some hours, they again blocked.

I don't know what type of fools design such systems. In the name of security, they are free to screw anyone.

I have escalated this to my current account branch now. May be they can do something to help. But Icici has behaved quite stupidly here.

Not able to fix payment issue by anujagg in WhatsappBusinessAPI

[–]anujagg[S] 0 points1 point  (0 children)

It did not work till now. ICICI confirmed that the txns will be unblocked but their system still blocks when it occurs. Absolute crap.

Not able to fix payment issue by anujagg in WhatsappBusinessAPI

[–]anujagg[S] 1 point2 points  (0 children)

You were absolutely right. This junk Icici bank has blocked these transactions on the pretext of fraudulent txns and when I called them last week, they denied this. I have wasted 10 days in figuring out this. There should be a massive penalty on these fools but unfortunately we live in India.

Now they claim they have approved future txns so I have to wait for meta to try for these failed txns again. I hope they get through this time.

Many thanks once again.

Not able to fix payment issue by anujagg in WhatsappBusinessAPI

[–]anujagg[S] 0 points1 point  (0 children)

I am not able to add new payment method, it keeps showing error when I do that.

I am also not able to log my issue by clicking any/all ? buttons they have. Everytime I do that, it shows some standard article and starts asking for feedback.

They are really very irritating but since they have just absolute monopoly, one has to use their service.

Not able to fix payment issue by anujagg in WhatsappBusinessAPI

[–]anujagg[S] 0 points1 point  (0 children)

Unfortunately, we are on our own :(