Skip to content

Commit c7f6dc3

Browse files
committed
fixes for removing prints and mock server
1 parent ae90cde commit c7f6dc3

File tree

4 files changed

+7
-8
lines changed

4 files changed

+7
-8
lines changed

src/guidellm/__main__.py

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@
4343
from guidellm.benchmark.scenario import (
4444
GenerativeTextScenario,
4545
)
46-
from guidellm.mock_server import MockServer, ServerConfig
46+
from guidellm.mock_server import MockServer, MockServerConfig
4747
from guidellm.preprocess.dataset import ShortPromptStrategy, process_dataset
4848
from guidellm.scheduler import StrategyType
4949
from guidellm.settings import print_config
@@ -658,6 +658,7 @@ def dataset(
658658
@click.option(
659659
"--model", default="llama-3.1-8b-instruct", help="The name of the model to mock"
660660
)
661+
@click.option("--processor", default=None, help="The processor to use for requests")
661662
@click.option(
662663
"--request-latency",
663664
default=3,
@@ -721,6 +722,7 @@ def mock_server(
721722
port: int,
722723
workers: int,
723724
model: str,
725+
processor: str | None,
724726
request_latency: float,
725727
request_latency_std: float,
726728
ttft_ms: float,
@@ -739,11 +741,12 @@ def mock_server(
739741
development scenarios requiring predictable server behavior.
740742
"""
741743

742-
config = ServerConfig(
744+
config = MockServerConfig(
743745
host=host,
744746
port=port,
745747
workers=workers,
746748
model=model,
749+
processor=processor,
747750
request_latency=request_latency,
748751
request_latency_std=request_latency_std,
749752
ttft_ms=ttft_ms,

src/guidellm/mock_server/__init__.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,7 @@
22
GuideLLM Mock Server for OpenAI and vLLM API compatibility.
33
"""
44

5+
from .config import MockServerConfig
56
from .server import MockServer
67

7-
__all__ = ["MockServer"]
8+
__all__ = ["MockServer", "MockServerConfig"]

src/guidellm/scheduler/worker.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -299,9 +299,6 @@ async def _process_next_request(self):
299299
request_info.scheduler_timings.resolve_end = time.time()
300300
self._send_update("completed", response, request, request_info)
301301

302-
print("\n\n********Completed request")
303-
print(request_info)
304-
305302
response = request = request_info = None
306303
except asyncio.CancelledError:
307304
# Handle cancellation

src/guidellm/scheduler/worker_group.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -479,7 +479,6 @@ def _iter():
479479
scheduler_start_time=self._start_time,
480480
)
481481
_, stop = self._locked_update(request_info, source="generator")
482-
print(f"----Sending request {request_info}")
483482
yield (request, request_info)
484483

485484
if stop:
@@ -517,7 +516,6 @@ def update_callback_receive(
517516
:return: Updated tuple with injected scheduler state
518517
"""
519518
response, request, request_info = update
520-
print(f"\n###########Received update for request: {request_info}")
521519
state, stop = self._locked_update(info=request_info, source="updates")
522520

523521
if stop:

0 commit comments

Comments
 (0)