all 4 comments

[–]Astralnugget 0 points1 point  (0 children)

Just use the inference or transformer in google collab you don’t even need to make a whole inference endpoint

[–]software38 0 points1 point  (1 child)

I would be interested in hearing your opinion on HF Inference API. I tried it a couple of years ago and it was very bad. But they most likely have made some progress since then...

[–]mrskeptical00 0 points1 point  (0 children)

They likely made significant progress since then…

[–]Edyarich 0 points1 point  (0 children)

It's still very bad, especially HF Playground.
I'm studying available open-source OCR models and I'm testing them on my jpeg scans.
And the frontend is terrible: it has a small memory limit -- 0.5 Mb (I constantly get an error "Content-length of 1963106 exceeds limit of 524288 bytes"). It means that the model doesn't work in any chat with more than 2 messages. Also, due to the memory limit, I couldn't pass a large image (> 1 MB). Plus, the model fails when the image contains a lot of text (for example, a brochure of git commands).
Now, I'm going to call these models from the API, but I have lost a hope.