Skip to content

Commit 432031e

Browse files
author
chibu
committed
test rag
1 parent 16b83bc commit 432031e

File tree

3 files changed

+17
-15
lines changed

3 files changed

+17
-15
lines changed

examples/guidellm_example.py

Lines changed: 0 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -11,22 +11,11 @@
1111
GUIDELLM__MAX_CONCURRENCY=256,
1212
GUIDELLM__REQUEST_TIMEOUT=21600,
1313
target="http://localhost:8000/v1",
14-
#target="http://fed73cc1-us-east.lb.appdomain.cloud/v1",
1514
data_type="emulated",
1615
max_seconds=30,
17-
#data="{'prompt_tokens': 512, 'generated_tokens': 256, 'output_tokens' : 256}",
1816
data="prompt_tokens=512,generated_tokens=256,output_tokens=256",
19-
#data="prompt_tokens=512,generated_tokens=256",
2017
branch = "update_guidellm",
2118
#vllm_kwargs={"enable-chunked-prefill": True}
2219
)
2320

24-
#from clearml import Task
25-
#task = Task.init(project_name="alexandre_debug", task_name="test_guidellm_task")
2621
task.execute_remotely("remote-upgrade-default")
27-
#task.execute_locally()
28-
import os
29-
import sys
30-
executable_path = os.path.dirname(sys.executable)
31-
vllm_path = os.path.join(executable_path, "vllm")
32-
print(f"The vllm path is: {vllm_path}")
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
{
2+
"rate_type": "sweep",
3+
"data": {
4+
"prompt_tokens": 4096,
5+
"prompt_tokens_stdev": 512,
6+
"prompt_tokens_min": 2048,
7+
"prompt_tokens_max": 6144,
8+
"output_tokens": 512,
9+
"output_tokens_stdev": 128,
10+
"output_tokens_min": 1,
11+
"output_tokens_max": 1024
12+
}
13+
}

src/automation/tasks/scripts/guidellm_script.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ def clean_hocon_value(v):
6363

6464
from pathlib import Path
6565
from guidellm.benchmark.scenario import GenerativeTextScenario, get_builtin_scenarios
66-
filepath = Path(os.path.join(".", "src", "automation", "standards", "benchmarking", "chat.json"))
66+
filepath = Path(os.path.join(".", "src", "automation", "standards", "benchmarking", "rag.json"))
6767
current_scenario = GenerativeTextScenario.from_file(filepath, dict(guidellm_args))
6868
# Start vLLM server
6969
server_process, server_initialized, server_log = start_vllm_server(
@@ -109,9 +109,9 @@ def clean_hocon_value(v):
109109
#default_scenario = get_builtin_scenarios()[0]
110110
#current_scenario = GenerativeTextScenario.from_builtin(default_scenario, dict(guidellm_args))
111111

112-
from pathlib import Path
113-
filepath = Path(os.path.join(".", "src", "automation", "standards", "benchmarking", "chat.json"))
114-
current_scenario = GenerativeTextScenario.from_file(filepath, dict(guidellm_args))
112+
#from pathlib import Path
113+
#filepath = Path(os.path.join(".", "src", "automation", "standards", "benchmarking", "chat.json"))
114+
#current_scenario = GenerativeTextScenario.from_file(filepath, dict(guidellm_args))
115115

116116
#import time
117117
#time.sleep(300)

0 commit comments

Comments
 (0)