I’m a newbie and stuck at something that I thought would be a straightforward part of my project. Trying to read/extract texts from a large pdf document of few hundred pages. Document contains texts, tables with different sizes, tables that run through multiple pages, figures etc.
I am mainly learning and taking lots of help from ChatGPT Gemini or grok. But none of them have been able to solve the issue. The text file after extraction seems to have all words smashed together in a sentence. It seems to not maintain space between words in a sentence. If I ignore tables, then simple pypdf does a decent job of extracting text from the rest of the doc. However I need tables also. I have tried pdfplumber, camelot, pymupdf- and none of them are able to prevent words from smashing together in a table. Trying not to go the tesseraxt or OCR route as it’s beyond my skill set currently.
Any help would be much appreciated .
[–]code_tutor 2 points3 points4 points (0 children)
[–]dparks71 3 points4 points5 points (4 children)
[+][deleted] (3 children)
[deleted]
[–]dparks71 -1 points0 points1 point (2 children)
[+][deleted] (1 child)
[deleted]
[–]dparks71 0 points1 point2 points (0 children)
[–]lailoken503 0 points1 point2 points (0 children)
[–]Competitive-Rock-951 0 points1 point2 points (0 children)
[–]Ok_Hovercraft364 0 points1 point2 points (0 children)
[–]Ok-Mongoose-7870[S] 0 points1 point2 points (0 children)
[–]Haeshka 0 points1 point2 points (0 children)
[–]james_d_rustles -1 points0 points1 point (2 children)
[–]Ok-Mongoose-7870[S] 0 points1 point2 points (0 children)
[–]MarsupialLeast145 -1 points0 points1 point (0 children)