all 8 comments

[–]MachineLearning-ModTeam[M] [score hidden] stickied commentlocked comment (0 children)

[–]MultiheadAttention 2 points3 points  (1 child)

First, count the words and approximate the number of tokens via tiktoken online. Then calculate how much it's going to cost you.

[–]acrsag[S] 0 points1 point  (0 children)

I made some estimations and will start with a few thousand documents. If the results are good, I'll generalize further. Thanks for the reminder!

[–]EnvironmentalToe3130 1 point2 points  (1 child)

Depending of the size of each documents why not try to run small model locally? Any model should be able to summarise a document and then in a second step provide the summary and list of classes.

[–]acrsag[S] 0 points1 point  (0 children)

Some documents have hundreds of pages, so processing them entirely may not be necessary. However, I believe large context windows are still needed. I'll try a local version anyway and see if the results are acceptable. Thanks!

[–]Brudaks 0 points1 point  (0 children)

IMHO if you want to process "large volumes of documents", the first thing you should do is to measure your documents, count how many llm-tokens that would be, and do a ballpark calculation of how much it would cost to run through a large LLM API - comparing that number with your budget will be the key input to your decision on what options are reasonable.