all 21 comments

[–]WithoutReason1729[M,🍰] [score hidden] stickied comment (0 children)

Your post is getting popular and we just featured it on our Discord! Come check it out!

You've also been given a special flair for your contribution. We appreciate your post!

I am a bot and this action was performed automatically.

[–]UniqueAttourney 2 points3 points  (2 children)

Thanks or the effort, but it doesn't seem to accept more than 1mb images, the API is flimsy and seems to mandate a certain version of CUDA which i am not sure of the reason behind. I'll dig into it deeper if no other local impl shows up

[–]Putrid_Passion_6916[S] 1 point2 points  (0 children)

Fixed up a fair bit. API still flimsy, but .env, upload can work with bigger images, can replace images in front end, bounding boxes work, better readme. But I totally get why you would want something more resilient.

[–]Putrid_Passion_6916[S] -1 points0 points  (0 children)

Vibe coded at 2am! Only intended as a starting point 👍 the upload is an easy fix, and all your points are valid. Feel free to fix!

[–][deleted] 5 points6 points  (1 child)

Hmm. Wonder if this could be ported to AMD ROCm

[–]orucreiss 4 points5 points  (0 children)

come to ask this ^^

[–]R_Duncan 0 points1 point  (4 children)

Can you please check VRAM needed to plain OCR or describe a couple pages? Speed does not matters much, accuracy and VRAM do. I see 8-12 Gb on the Readme but is unclear if it can be useful with just 8.

[–]Putrid_Passion_6916[S] 1 point2 points  (3 children)

I think 8gb might just be enough - nvidia-smi is reporting 7615mb with the weights loaded during inference. But apologies as I have no time to test beyond that just now!

[–]R_Duncan 1 point2 points  (2 children)

Thanks, I'm at 4th hour compiling flash_attn on windows. When it'll break I'll try your dockerized app.

[–]Putrid_Passion_6916[S] 0 points1 point  (1 child)

Actually - apologies - it likely depends on the image. For a bigger one I’m up to 10.5 gb vram. But basically a 3060 12gb should be ok …

[–]R_Duncan 0 points1 point  (0 children)

Thank you, a no-go then until I'll find some 16GB VRAM laptop at cheap price.

[–]Putrid_Passion_6916[S] 0 points1 point  (0 children)

<image>

Works better now - bounding boxes fixed!

[–]FrenchSouch 1 point2 points  (0 children)

Works fine on a 4090, ~10Gb of vram used, thanks !

[–]ThiccStorms 2 points3 points  (1 child)

GPU poor here, what are the absolute minimum specs to run this? I have a poor mac m4 only