You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: model_api/python/model_api/adapters/ovms_adapter.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,15 +7,15 @@ The `OVMSAdapter` implements `InferenceAdapter` interface. The `OVMSAdapter` mak
7
7
`OVMSAdapter` enables inference via gRPC calls to OpenVINO Model Server, so in order to use it you need two things:
8
8
9
9
- OpenVINO Model Server that serves your model
10
-
-[`ovmsclient`](https://pypi.org/project/ovmsclient/) package installed to enable communication with the model server: `python3 -m pip install ovmsclient`
10
+
-[`tritonclient[http]`](https://pypi.org/project/tritonclient/) package installed to enable communication with the model server: `python3 -m pip install tritonclient[http]`
11
11
12
12
### Deploy OpenVINO Model Server
13
13
14
14
Model Server is distributed as a docker image and it's available in DockerHub, so you can use it with `docker run` command. See [model server documentation](https://github.com/openvinotoolkit/model_server/blob/main/docs/starting_server.md) to learn how to deploy OpenVINO optimized models with OpenVINO Model Server.
15
15
16
16
## Model configuration
17
17
18
-
When using OpenVINO Model Server model cannot be directly accessed from the client application (like OMZ demos). Therefore any configuration must be done on model server side or before starting the server: see [Prepare a model for `InferenceAdapter`](../../../../../README.md#prepare-a-model-for-inferenceadapter).
18
+
When using OpenVINO Model Server model cannot be directly accessed from the client application. Therefore any configuration must be done on model server side or before starting the server: see [Prepare a model for `InferenceAdapter`](../../../../../README.md#prepare-a-model-for-inferenceadapter).
19
19
20
20
### Input reshaping
21
21
@@ -51,8 +51,8 @@ To run the demo with model served in OpenVINO Model Server, you would have to pr
51
51
52
52
Assuming that model server runs on the same machine as the demo, exposes gRPC service on port 9000 and serves model called `model1`, the value of `-m` parameter would be:
53
53
54
-
-`localhost:9000/models/model1` - requesting latest model version
55
-
-`localhost:9000/models/model1:2` - requesting model version number 2
54
+
-`localhost:9000/v2/models/model1` - requesting latest model version
55
+
-`localhost:9000/v2/models/model1:2` - requesting model version number 2
0 commit comments