You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Note that the provider name is needed to deploy provider-specific resources. If no provider is specified, then only the InferencePool object and the EPP are deployed.
24
24
25
+
### Install with Custom Environment Variables
26
+
27
+
To set custom environment variables for the EndpointPicker deployment:
Use `--set inferencePool.modelServerType=triton-tensorrt-llm` to install for Triton TensorRT-LLM, e.g.,
@@ -57,6 +84,7 @@ The following table list the configurable parameters of the chart.
57
84
|`inferenceExtension.image.tag`| Image tag of the endpoint picker. |
58
85
|`inferenceExtension.image.pullPolicy`| Image pull policy for the container. Possible values: `Always`, `IfNotPresent`, or `Never`. Defaults to `Always`. |
59
86
|`inferenceExtension.extProcPort`| Port where the endpoint picker service is served for external processing. Defaults to `9002`. |
87
+
|`inferenceExtension.env`| Map of environment variables to set in the endpoint picker container. Defaults to `{}`. |
60
88
|`provider.name`| Name of the Inference Gateway implementation being used. Possible values: `gke`. Defaults to `none`. |
0 commit comments