Skip to content

Commit 768d135

Browse files
author
chibu
committed
back to base
1 parent 629d195 commit 768d135

File tree

4 files changed

+24
-11
lines changed

4 files changed

+24
-11
lines changed

setup.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,9 +17,6 @@
1717
#"google-cloud-storage>=1.13.2",
1818
"datasets",
1919
"pyhocon",
20-
"build>=1.0.0",
21-
"setuptools>=61.0",
22-
"setuptools-git-versioning>=2.0,<3",
2320
],
2421
python_requires=">=3.7",
2522
)
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
{
2+
"rate_type": "sweep",
3+
"data": {
4+
"prompt_tokens": 512,
5+
"prompt_tokens_stdev": 128,
6+
"prompt_tokens_min": 1,
7+
"prompt_tokens_max": 1024,
8+
"output_tokens": 256,
9+
"output_tokens_stdev": 64,
10+
"output_tokens_min": 1,
11+
"output_tokens_max": 1024
12+
}
13+
}

src/automation/tasks/guidellm.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,8 @@
44
import os
55

66
DEFAULT_SERVER_WAIT_TIME = 600 # 600 seconds = 10 minutes
7-
GUIDELLM_PACKAGE = "git+https://github.com/neuralmagic/guidellm.git@clearml-guidellm#egg=guidellm"
7+
GUIDELLM_PACKAGE = "git+https://github.com/neuralmagic/guidellm.git"
8+
#GUIDELLM_PACKAGE = "git+https://github.com/neuralmagic/guidellm.git@clearml-guidellm#egg=guidellm"
89
#GUIDELLM_PACKAGE = "git+https://github.com/neuralmagic/guidellm.git@clearml-guidellm"
910
#GUIDELLM_PACKAGE = "git+https://github.com/neuralmagic/guidellm.git@main#egg=guidellm[dev]"
1011

@@ -14,7 +15,7 @@ class GuideLLMTask(BaseTask):
1415
#"build>=1.0.0",
1516
#"setuptools>=61.0",
1617
#"setuptools-git-versioning>=2.0,<3",
17-
#"vllm",
18+
"vllm",
1819
GUIDELLM_PACKAGE,
1920
"hf_xet",
2021
]

src/automation/tasks/scripts/guidellm_script.py

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -61,9 +61,9 @@ def clean_hocon_value(v):
6161
print(gpu_count)
6262
print(os.getcwd())
6363

64-
from guidellm.benchmark.scenario import GenerativeTextScenario, get_builtin_scenarios
65-
print(get_builtin_scenarios())
66-
default_scenario = get_builtin_scenarios()[0]
64+
from pathlib import Path
65+
filepath = Path(os.path.join(".", "src", "automation", "standards", "benchmarking", "chat.json"))
66+
current_scenario = GenerativeTextScenario.from_file(filepath, dict(guidellm_args))
6767
# Start vLLM server
6868
server_process, server_initialized, server_log = start_vllm_server(
6969
vllm_args,
@@ -105,10 +105,12 @@ def clean_hocon_value(v):
105105
print(f"The vllm path is: {vllm_path}")
106106

107107

108-
print(get_builtin_scenarios())
109-
default_scenario = get_builtin_scenarios()[0]
108+
#default_scenario = get_builtin_scenarios()[0]
109+
#current_scenario = GenerativeTextScenario.from_builtin(default_scenario, dict(guidellm_args))
110110

111-
current_scenario = GenerativeTextScenario.from_builtin(default_scenario, dict(guidellm_args))
111+
from pathlib import Path
112+
filepath = Path(os.path.join(".", "src", "automation", "standards", "benchmarking", "chat.json"))
113+
current_scenario = GenerativeTextScenario.from_file(filepath, dict(guidellm_args))
112114

113115
#import time
114116
#time.sleep(300)

0 commit comments

Comments
 (0)