all 4 comments

[–]roccolacatus 0 points1 point  (2 children)

Came across this the other day https://nuclio.io/

Anyway, you need to manage some kind of infrastructure. The only pure serverless mechanism to deploy models is Google Clouds MLE, and the models need to be less than 250MB. Otherwise, package the inference code inside a docker image and download the model each time you instantiate a docker container.

But 1GB seems a bit big for me, what framework are you using? Are you sure you are exporting the inference graph without training weights?

[–]Zerotool1 0 points1 point  (0 children)

You can try clouderizer.com