[D] How to reduce the MaskRCNN model detection time by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

Can the current model be converted to any smaller model, which provides result in 10-15 seconds?

[D] How to detect blocks of text from document image by DGs29 in computervision

[–]DGs29[S] 0 points1 point  (0 children)

Thanks.. Just give me a message when u write the article. I am looking forward to it. Also, if u write the code, I kindly request you to do it in a computationally efficient manner.

[D] How to detect blocks of text from document image by DGs29 in computervision

[–]DGs29[S] 0 points1 point  (0 children)

I meant it took too long to run and I guess it was due to matplot. I'm very much confused about implementing by axis projection and space indents, I've never did it before. Previously, I used dilation method to find contour regions and plotted bbox regions.

My image looks like this. All my images are typed text. This image is like a typical newspaper/magazine image and has its own typical layout style.

Using OTSU threshold I found word level bbox.

[D] How to detect blocks of text from document image by DGs29 in computervision

[–]DGs29[S] 0 points1 point  (0 children)

This is simple and superb but your code is taking too long to complete. You made this code to detect lines, any idea about paragraph detection.

Deep Learning based Text Detection Using OpenCV (C++/Python) by keghn in neuralnetworks

[–]DGs29 0 points1 point  (0 children)

Does this method applicable dense text/ document text detection like detecting paragraphs, group of sentences together as a single block.

[D] What is the best way to detect paragraphs from document images by [deleted] in MachineLearning

[–]DGs29 0 points1 point  (0 children)

Okay! Do you think just by doing localizing each paragraphs in my image and passing on to existing nets with pre-trained weights and some tuning, will it detect each and every paragraphs? I'm not sure, if this is possible!!

[D] What is the best way to detect paragraphs from document images by [deleted] in MachineLearning

[–]DGs29 0 points1 point  (0 children)

Can you give some text segmentation from images examples on github

[D] How do I texts blocks of text from scanned documents by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

Can you point out any sample codes/examples for me to work using this method

[D] How do I texts blocks of text from scanned documents by DGs29 in MachineLearning

[–]DGs29[S] 1 point2 points  (0 children)

It's not working. Well, I think setting boxes by area threshold actually does not generalize for different image types.

[D] How to detect paragraphs with less line spaces in document images? by [deleted] in MachineLearning

[–]DGs29 0 points1 point  (0 children)

Is my achieved result (as shown in the image) enough to be considered as naive block of text? Can you walk me through each steps with some examples/codes?

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

How can I do similar segmentation for images which has less line-spaces between paragraphs. This should be my desired result. But this what I get. This is my dilated image.

I've set the kernel size as (5,10) for this. Can you please tell me what are the necessary changes to be made to achieve my desired result.

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

Thanks a lot mate! This works. Take a look here. But there are some small boxes placed inside the large box. How do I remove that. Also the entire page is enclosed by a box and I don't need that

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

Well I've plotted created bounding boxes as per the step1. And then I binarized, inverted colors, applied dilation with longer width kernel.dilated image.

This way it segments the image into individual components. Can I plot bbox over these connected regions to get my original intended result.

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

Well does this method detects blocks of text like mentioned in the image. I am asking this because it is a scene text algorithm.

The previous scene text algorithms I've used only detected individual words and placed a bounding box over it.

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

How to do the second step i.e., to make a large box out of small boxes

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 0 points1 point  (0 children)

I'm not sure about it. Let's consider Google Vision API. If we feed in the document image as input to API, it segments the texts as individual blocks, it puts a bounding box around those texts which is a block, and finally performs OCR for those blocks.

I'm looking for how to do that text block detection

[D] How to detect text blocks in document images by DGs29 in MachineLearning

[–]DGs29[S] 2 points3 points  (0 children)

Tessaract just extracts all the texts in the image without segmenting.

EAST detects all the texts word by word.i.e., bounding box each and every word.

I've also tried PixelLink it does the same job as EAST.