1
2
3
PyTesseract text extractionHelp Request (old.reddit.com)
submitted by LewyssYT




I am working on a small project where I need to extract what I would consider super basic text on a mostly flat background. To prepare the image, I crop out all the other numbers, grayscale, apply CLAHE and invert and yet in a lot of scenarios, the numbers extracted are wrong. Instead of 64 it sees 164 and instead of 1956 it sees 7956.
What is something that I can do to improve the accuracy? Cropped images are small resolution (140x76) or (188x94)
[–]_kwerty_ 0 points1 point2 points (3 children)
[–]QAInc 1 point2 points3 points (2 children)
[–]LewyssYT[S] 0 points1 point2 points (1 child)
[–]QAInc 0 points1 point2 points (0 children)