Skip to content

Commit 2ed0d81

Browse files
committed
Add arguments for updated vllm service
1 parent 79f67ea commit 2ed0d81

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

llm-vllm-deployer/steps/vllm_deployer.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -31,6 +31,7 @@
3131
@step(enable_cache=False)
3232
def vllm_model_deployer_step(
3333
model: str,
34+
port: int = 8000,
3435
tokenizer: Optional[str] = None,
3536
timeout: int = 1200,
3637
deploy_decision: bool = True,
@@ -41,6 +42,7 @@ def vllm_model_deployer_step(
4142
4243
Args:
4344
model: Name or path to huggingface model
45+
port: Port used by vllm server
4446
tokenizer: Name or path of the huggingface tokenizer to use.
4547
If unspecified, model name or path will be used.
4648
timeout: the number of seconds to wait for the service to start/stop.
@@ -62,6 +64,7 @@ def vllm_model_deployer_step(
6264
# create a config for the new model service
6365
predictor_cfg = VLLMServiceConfig(
6466
model=model,
67+
port=port,
6568
tokenizer=tokenizer,
6669
model_name="default", # Required for ServiceConfig
6770
)

0 commit comments

Comments
 (0)