- Your ~/.kube/config should point to a cluster with KFServing installed.
- Your cluster's Istio Ingress gateway must be network accessible.
Apply the CRD
kubectl apply -f onnx.yaml
Expected Output
$ inferenceservice.serving.kubeflow.org/style-sample configured
- Setup env vars
The first step is to determine the ingress IP and ports and set
INGRESS_HOST
andINGRESS_PORT
export MODEL_NAME=style-sample
export SERVICE_HOSTNAME=$(kubectl get inferenceservice ${MODEL_NAME} -o jsonpath='{.status.url}' | cut -d "/" -f 3)
- Verify the service is healthy
SERVICE_URL=http://${INGRESS_HOST}:${INGRESS_PORT}/v1/models/$MODEL_NAME
curl ${SERVICE_URL}
- Install dependencies
pip install -r requirements.txt
- Run the sample notebook in jupyter
jupyter notebook
The sample model for the example in this readme is already uploaded and available for use. However if you would like to modify the example to use your own ONNX model, all you need to do is to upload your model as model.onnx
to S3, GCS or an Azure Blob.