This repository was archived by the owner on Nov 16, 2023. It is now read-only.
File tree Expand file tree Collapse file tree 1 file changed +6
-1
lines changed
Research/kubeflow-on-azure-stack-lab/04-KFServing/triton Expand file tree Collapse file tree 1 file changed +6
-1
lines changed Original file line number Diff line number Diff line change @@ -84,6 +84,11 @@ And let us set the following:
8484 $ INPUT_PATH=@./triton_input.json
8585 $ SERVICE_HOSTNAME=$(kubectl get inferenceservices -n kfserving-test bert-large -o jsonpath='{.status.url}' | cut -d "/" -f 3)
8686
87+ You should see the service ` Alive ` :
88+
89+ $ curl -H "Host: ${SERVICE_HOSTNAME}" http://${INGRESS_HOST}:${INGRESS_PORT}
90+ Alive
91+
8792And now we can run the web requests for the inferencing server we created:
8893
8994 $ curl -v -H "Host: ${SERVICE_HOSTNAME}" -d $INPUT_PATH http://${INGRESS_HOST}:${INGRESS_PORT}/v1/models/$MODEL_NAME:predict
@@ -104,7 +109,7 @@ And now we can run the web requests for the inferencing server we created:
104109 < server: istio-envoy
105110 < x-envoy-upstream-service-time: 3814
106111 <
107- * Connection #0 to host 38.102.181.75 left intact
112+ * Connection #0 to host 12.34.56.78 left intact
108113 {"predictions": "John F. Kennedy", "prob": 77.91852121017916}
109114
110115So, we got JFK with 78% certainty, which is reasonable.
You can’t perform that action at this time.
0 commit comments