Seldon Serving

Model serving using Seldon

Serve a model using Seldon

Seldon comes installed with Kubeflow. Full documentation for running Seldon inference is provided within the Seldon documentation site.

If you have a saved model in a PersistentVolume (PV), Google Cloud Storage bucket or Amazon S3 Storage you can use one of the prepackaged model servers provided by Seldon.

Seldon also provides language specific model wrappers to wrap your inference code for it to run in Seldon.

Kubeflow Specifics

Examples

Seldon provides a large set of example notebooks showing how to run inference code for a wide range of machine learning toolkits.