Skip to content

Commit 6a91524

Browse files
[https://nvbugs/5648685][fix] Fix openAI server waiting time to avoid large model weight loading out time (#9254)
Signed-off-by: Wangshanshan <30051912+dominicshanshan@users.noreply.github.com> Signed-off-by: Mike Iovine <6158008+mikeiovine@users.noreply.github.com> Signed-off-by: Mike Iovine <miovine@nvidia.com> Signed-off-by: Wangshanshan <30051912+dominicshanshan@users.noreply.github.com>
1 parent d5e6b67 commit 6a91524

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

tests/unittest/llmapi/apps/openai_server.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@
1616

1717
class RemoteOpenAIServer:
1818
DUMMY_API_KEY = "tensorrt_llm"
19-
MAX_SERVER_START_WAIT_S = 600 # wait for server to start for 600 seconds
19+
MAX_SERVER_START_WAIT_S = 7200 # wait for server to start for 7200 seconds (~ 2 hours) for LLM models weight loading
2020

2121
def __init__(self,
2222
model: str,

0 commit comments

Comments
 (0)