File tree Expand file tree Collapse file tree 2 files changed +4
-6
lines changed Expand file tree Collapse file tree 2 files changed +4
-6
lines changed Original file line number Diff line number Diff line change 1
-
2
1
from automation .tasks import GuideLLMTask
3
2
4
3
task = GuideLLMTask (
5
4
project_name = "alexandre_debug" ,
6
5
task_name = "test_guidellm_task" ,
7
- #model="meta-llama/Llama-3.2-1B-Instruct",
8
- model = "Qwen/Qwen2.5-1.5B-Instruct" ,
6
+ model = "meta-llama/Llama-3.2-1B-Instruct" ,
9
7
rate_type = "throughput" ,
10
8
backend = "aiohttp_server" ,
11
9
GUIDELLM__MAX_CONCURRENCY = 256 ,
12
10
GUIDELLM__REQUEST_TIMEOUT = 21600 ,
13
11
target = "http://localhost:8000/v1" ,
14
12
max_seconds = 30 ,
15
- scenario = "benchmarking_32kz " ,
13
+ # scenario = "benchmarking_32k ",
16
14
data = "prompt_tokens=128,output_tokens=128" ,
17
15
branch = "update_guidellm" ,
18
16
vllm_kwargs = {"enable-chunked-prefill" : True }
19
17
)
20
18
21
- task .execute_remotely ("remote-upgrade-default" )
19
+ task .execute_remotely ("oneshot-a100x1" )
20
+ #task.execute_locally()
Original file line number Diff line number Diff line change 1
1
DEFAULT_DOCKER_IMAGE = "498127099666.dkr.ecr.us-east-1.amazonaws.com/mlops/k8s-research-cuda12_8:latest"
2
2
DEFAULT_OUTPUT_URI = "gs://neuralmagic-clearml"
3
- #DEFAULT_OUTPUT_URI = "http://10.128.20.60:8081"
4
3
DEFAULT_RESEARCH_BRANCH = "main"
5
4
DEFAULT_GUIDELLM_SCENARIO = "chat"
You can’t perform that action at this time.
0 commit comments