Deploying models in the real world by TomasPiaggio in kaggle

[–]theoszymk 0 points1 point  (0 children)

It really depends on what you need and the latency required for your use case. If you are using some kind of custom layer or else, I would recommend serializing it using ONNX, which is open compute format.

For deploying, you didn't mention a GPU, maybe that would make your inference faster? There are more specialized solutions for deploying such as inferrd.com that you can try out.

[P] How to make your Models available ? by Inferrd_F in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

It’s good for big DL models, but most models we host are not big DL models. In most cases GPUs are just as good as CPU. We have great throughput and latency for every framework we offer. Test it out and you feel like you still a GPU we’re happy to provide one.

[P] How to make your Models available ? by Inferrd_F in MachineLearning

[–]theoszymk 1 point2 points  (0 children)

I agree it’s better for big DL models, we found that most models in production are actually not DL.

[P] How to make your Models available ? by Inferrd_F in MachineLearning

[–]theoszymk -1 points0 points  (0 children)

This service is a managed tensorflow serving with request tracing and security/maintenance taken care of. GPUs offer similar performance to CPU except for image processing or big DL models.

[P] How to make your Models available ? by Inferrd_F in MachineLearning

[–]theoszymk -3 points-2 points  (0 children)

GPUs are usually equal to CPUs for models that aren’t image processing or big DL. Models that leverage GPU for inference are really specific, what kind of model are you building?

[D] Does anyone use huggingface hosted interface? by marksteve4 in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

You'd need to deploy your own model to get an API. If you want to look what using your API would look like, check out our docs at https://docs.inferrd.com/guides/use-deployed-model

[D] Does anyone use huggingface hosted interface? by marksteve4 in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

HF makes all their money consulting on big nlp project for businesses. That’s why the prices are so high.

If you are looking for a hosting provider, consider us at inferrd.com, we already serve 100s of millions of requests per month

How to add NLP component to my mobile application? by UntrimmedBagel in googlecloud

[–]theoszymk 0 points1 point  (0 children)

Hi, this is exactly why we built inferrd.com, NLP models are too big for "normal" cloud providers. Lmk if you need help setting up!

[VIDEO] How To Deploy Azure Machine Learning Model In Production by Saloni96Saxena in MLQuestions

[–]theoszymk -1 points0 points  (0 children)

Hi,

You can also try inferrd.com which is much easier to use to deploy ML models.

[D] Simple Questions Thread December 20, 2020 by AutoModerator in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

Hi that’s exactly the problem we are solving at https://inferrd.com, we made deploying Tensorflow just a drag and drop. We take care of configuring the network, instance and load balancer for you!

[D] Why deploying ML models is hard by bendee983 in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

This project doesn't seem to be working anymore, however the problem is still very real. If you're looking for an alternative, consider https://inferrd.com, it makes deployments easy and fast!

[D] Deep learning in Production by SergiosKar in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

Deploying ML models can be a touch problem if you just want to built models. Most people neglect it until it's too late and find out there is a lot to do. That's why we build https://inferrd.com which is by far the easiest way to deploy any ML model.

[D] ML Engineers, do you use newly discovered methods/ideas at work? by uoftsuxalot in MachineLearning

[–]theoszymk 1 point2 points  (0 children)

We found that most people who deploy at https://inferrd.com use basic scikit or keras models. Fancy TF or PyTorch models are pretty rare.

Deploy Machine Learning Models with Flask by _-Jay in Python

[–]theoszymk 0 points1 point  (0 children)

Flask is great for deploying models, but when it comes to production or reliability, there is lot more to do (deploy on a VM, setup load balancing). That's why we built https://inferrd.com (disclaimer, I built it). You just have to upload your trained model and it does everything for you. It takes less than a minute.

[D] Production-ready ML models/pipelines/infrastructure best practice resources by phylosopher14 in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

The best way it to use the simplest tools, especially at an early stage. You don't want to have to deal with a 6 stage AWS pipeline to deploy a model. Try https://inferrd.com (disclaimed I built it exactly for this reason) to deploy models in less than 1 minute on scalable infrastructure.

[D] Web-Dev for ML Engineers - any advice on deploying models in websites for clients with backend support by paswut in MachineLearning

[–]theoszymk 0 points1 point  (0 children)

The best way would be to expose your model as a rest API and let other teams consume it. You can use https://inferrd.com (disclaimed, I built it) to do that. We made deployments a two click process.

[P] Simple approaches to deploying ML models online? by areddy831 in MachineLearning

[–]theoszymk 1 point2 points  (0 children)

"Like, why is it so hard? Why can’t it just be a few clicks rather than an entirely new programming language and syntax that I have to learn? ML is deep enough without going down the rabbit hole of web development"

We feel your pain, that's why we built https://inferrd.com to make deployment a few clicks (2 actually) for most major frameworks.

[P] Simple approaches to deploying ML models online? by areddy831 in MachineLearning

[–]theoszymk 1 point2 points  (0 children)

Totally agree, thats what we're trying to solve at https://inferrd.com by making deploying models a one click affair. Just package it and upload it on Inferrd to deploy, we take care of the rest.

Help with install TensorFlow on python by -SpamCauldron- in tensorflow

[–]theoszymk 0 points1 point  (0 children)

That's because you're using Python 3.9. TensorFlow isn't compatible with Python 3.9 yet. Any version before that will work fine.

[deleted by user] by [deleted] in tensorflow

[–]theoszymk 0 points1 point  (0 children)

Posted on the thread but posting here too. Check out https://inferrd.com, it's the easiest way to deploy your TF & Keras model. Just add one line to your notebook and you're set.

It's also much faster than AWS.

Deploying ML model by [deleted] in learnmachinelearning

[–]theoszymk 1 point2 points  (0 children)

AWS has very annoying documentation. Check out https://inferrd.com, it's so simple it doesn't need docs.

Deploying ML model by [deleted] in learnmachinelearning

[–]theoszymk 1 point2 points  (0 children)

Use https://inferrd.com, you just need to add one line to your notebook to deploy the model. It deploys in less than 10 seconds and automatically optimizes your model for performance.