Version v0.6 of the documentation is no longer actively maintained. The site that you are currently viewing is an archived snapshot. For up-to-date documentation, see the latest version.

Serving

Serving of ML models in Kubeflow

Istio Integration (for TF Serving)

Using Istio for TF Serving

Seldon Serving

Model serving using Seldon

NVIDIA TensorRT Inference Server

Model serving using TRT Inference Server

TensorFlow Serving

Serving TensorFlow models

TensorFlow Batch Predict

Batch prediction for TensorFlow models

PyTorch Serving

Instructions for serving a PyTorch model with Seldon