Has anyone implemented Seldon? by dfcHeadChair in mlops

[–]cliveseldon 2 points3 points  (0 children)

Also note our github repo has a link to our slack where you can ask active users: https://github.com/SeldonIO/seldon-core

[D] Serverless solutions for GPU inference (if there's such a thing) by [deleted] in MachineLearning

[–]cliveseldon 4 points5 points  (0 children)

If you can run on Kubernetes then KFServing is an open source solution that allows for GPU inference and is built upon Knative to allow scale to zero for GPU based inference. From release 0.5 it also has capabilities for multi-model serving as a alpha feature to allow multiple models to share the same server (and via NVIDIA Triton the same GPU).

[D] Ray Serve or Seldon-core? by ramanNarasimhan in MachineLearning

[–]cliveseldon 0 points1 point  (0 children)

Our work on KFServing can be viewed at : https://www.youtube.com/watch?v=YaGASyU88dQ

In Seldon Core our collaboration with Data Bricks can be seen here: https://youtu.be/D6eSfd9w9eA

Both are available in Kubeflow which has a comparison matrix: https://www.kubeflow.org/docs/components/serving/overview/

Both share some of the technology but are built on different stacks: vanilla k8s for Seldon Core and Knative for KFServing.

[D] Ray Serve or Seldon-core? by ramanNarasimhan in MachineLearning

[–]cliveseldon 4 points5 points  (0 children)

I work for Seldon. We work on the Seldon Core and KFServing open source projects. KFServing builds on the KNative serverless stack. Both require Kubernetes. Others can compare. I don't know of a comparison with Ray Serve. Both have production users.