To help with development and testing we have developed a light weight vLLM simulator. It does not truly run inference, but it does emulate responses to the HTTP REST endpoints of vLLM. Currently it supports partial OpenAI-compatible API:
- /v1/chat/completions
- /v1/completions
- /v1/models
In addition, a set of the vLLM HTTP endpoints are suppored as well. These include:
| Endpoint | Description | 
|---|---|
| /v1/load_lora_adapter | simulates the dynamic registration of a LoRA adapter | 
| /v1/unload_lora_adapter | simulates the dynamic unloading and unregistration of a LoRA adapter | 
| /metrics | exposes Prometheus metrics. See the table below for details | 
| /health | standard health check endpoint | 
| /ready | standard readiness endpoint | 
In addition, it supports a subset of vLLM's Prometheus metrics. These metrics are exposed via the /metrics HTTP REST endpoint. Currently supported are the following metrics:
| Metric | Description | 
|---|---|
| vllm:gpu_cache_usage_perc | The fraction of KV-cache blocks currently in use (from 0 to 1). Currently this value will always be zero. | 
| vllm:lora_requests_info | Running stats on LoRA requests | 
| vllm:num_requests_running | Number of requests currently running on GPU | 
| vllm:num_requests_waiting | Prometheus metric for the number of queued requests | 
The simulated inference has no connection with the model and LoRA adapters specified in the command line parameters or via the /v1/load_lora_adapter HTTP REST endpoint. The /v1/models endpoint returns simulated results based on those same command line parameters and those loaded via the /v1/load_lora_adapter HTTP REST endpoint.
The simulator supports two modes of operation:
- echomode: the response contains the same text that was received in the request. For- /v1/chat/completionsthe last message for the role=- useris used.
- randommode: the response is randomly chosen from a set of pre-defined sentences.
Timing of the response is defined by the time-to-first-token and inter-token-latency parameters. In case P/D is enabled for a request, kv-cache-transfer-latency will be used instead of time-to-first-token.
For a request with stream=true: time-to-first-token or kv-cache-transfer-latency defines the delay before the first token is returned, inter-token-latency defines the delay between subsequent tokens in the stream.
For a requst with stream=false: the response is returned after delay of <time-to-first-token> + (<inter-token-latency> * (<number_of_output_tokens> - 1)) or <kv-cache-transfer-latency> + (<inter-token-latency> * (<number_of_output_tokens> - 1)) in P/D case
It can be run standalone or in a Pod for testing under packages such as Kind.
API responses contains a subset of the fields provided by the OpenAI API.
Click to show the structure of requests/responses
- /v1/chat/completions- request
- stream
- model
- messages
- role
- content
 
 
- response
- id
- created
- model
- choices
- index
- finish_reason
- message
 
 
 
- request
- /v1/completions- request
- stream
- model
- prompt
- max_tokens (for future usage)
 
- response
- id
- created
- model
- choices
- text
 
 
 
- request
- /v1/models- response
- object (list)
- data
- id
- object (model)
- created
- owned_by
- root
- parent
 
 
 
- response
For more details see the vLLM documentation
- config: the path to a yaml configuration file that can contain the simulator's command line parameters. If a parameter is defined in both the config file and the command line, the command line value overwrites the configuration file value. An example configuration file can be found at- manifests/config.yaml
- port: the port the simulator listents on, default is 8000
- model: the currently 'loaded' model, mandatory
- served-model-name: model names exposed by the API (a list of space-separated strings)
- lora-modules: a list of LoRA adapters (a list of space-separated JSON strings): '{"name": "name", "path": "lora_path", "base_model_name": "id"}', optional, empty by default
- max-loras: maximum number of LoRAs in a single batch, optional, default is one
- max-cpu-loras: maximum number of LoRAs to store in CPU memory, optional, must be >= than max-loras, default is max-loras
- max-model-len: model's context window, maximum number of tokens in a single request including input and output, optional, default is 1024
- max-num-seqs: maximum number of sequences per iteration (maximum number of inference requests that could be processed at the same time), default is 5
- max-waiting-queue-length: maximum length of inference requests waiting queue, default is 1000
- mode: the simulator mode, optional, by default- random- echo: returns the same text that was sent in the request
- random: returns a sentence chosen at random from a set of pre-defined sentences
 
- time-to-first-token: the time to the first token (in milliseconds), optional, by default zero
- time-to-first-token-std-dev: standard deviation for time before the first token will be returned, in milliseconds, optional, default is 0, can't be more than 30% of- time-to-first-token, will not cause the actual time to first token to differ by more than 70% from- time-to-first-token
- inter-token-latency: the time to 'generate' each additional token (in milliseconds), optional, by default zero
- inter-token-latency-std-dev: standard deviation for time between generated tokens, in milliseconds, optional, default is 0, can't be more than 30% of- inter-token-latency, will not cause the actual inter token latency to differ by more than 70% from- inter-token-latency
- kv-cache-transfer-latency: time for KV-cache transfer from a remote vLLM (in milliseconds), by default zero. Usually much shorter than- time-to-first-token
- kv-cache-transfer-latency-std-dev: standard deviation for time to "transfer" kv-cache from another vLLM instance in case P/D is activated, in milliseconds, optional, default is 0, can't be more than 30% of- kv-cache-transfer-latency, will not cause the actual latency to differ by more than 70% from- kv-cache-transfer-latency
- prefill-overhead: constant overhead time for prefill (in milliseconds), optional, by default zero, used in calculating time to first token, this will be ignored if- time-to-first-tokenis not- 0
- prefill-time-per-token: time taken to generate each token during prefill (in milliseconds), optional, by default zero, this will be ignored if- time-to-first-tokenis not- 0
- prefill-time-std-dev: similar to- time-to-first-token-std-dev, but is applied on the final prefill time, which is calculated by- prefill-overhead,- prefill-time-per-token, and number of prompt tokens, this will be ignored if- time-to-first-tokenis not- 0
- kv-cache-transfer-time-per-token: time taken to transfer cache for each token in case P/D is enabled (in milliseconds), optional, by default zero, this will be ignored if- kv-cache-transfer-latencyis not- 0
- kv-cache-transfer-time-std-dev: similar to- time-to-first-token-std-dev, but is applied on the final kv cache transfer time in case P/D is enabled (in milliseconds), which is calculated by- kv-cache-transfer-time-per-tokenand number of prompt tokens, this will be ignored if- kv-cache-transfer-latencyis not- 0
- time-factor-under-load: a multiplicative factor that affects the overall time taken for requests when parallelrequests are being processed. The value of this factor must be >= 1.0, with a default of 1.0. If this factor is 1.0, no extra time is added. When the factor is x (where x > 1.0) and there are- max-num-seqsrequests, the total time will be multiplied by x. The extra time then decreases multiplicatively to 1.0 when the number of requests is less than MaxNumSeqs.
- seed: random seed for operations (if not set, current Unix time in nanoseconds is used)
- max-tool-call-integer-param: the maximum possible value of integer parameters in a tool call, optional, defaults to 100
- min-tool-call-integer-param: the minimum possible value of integer parameters in a tool call, optional, defaults to 0
- max-tool-call-number-param: the maximum possible value of number (float) parameters in a tool call, optional, defaults to 100
- min-tool-call-number-param: the minimum possible value of number (float) parameters in a tool call, optional, defaults to 0
- max-tool-call-array-param-length: the maximum possible length of array parameters in a tool call, optional, defaults to 5
- min-tool-call-array-param-length: the minimum possible length of array parameters in a tool call, optional, defaults to 1
- tool-call-not-required-param-probability: the probability to add a parameter, that is not required, in a tool call, optional, defaults to 50
- object-tool-call-not-required-field-probability: the probability to add a field, that is not required, in an object in a tool call, optional, defaults to 50
- enable-kvcache: if true, the KV cache support will be enabled in the simulator. In this case, the KV cache will be simulated, and ZQM events will be published when a KV cache block is added or evicted.
- kv-cache-size: the maximum number of token blocks in kv cache
- block-size: token block size for contiguous chunks of tokens, possible values: 8,16,32,64,128
- tokenizers-cache-dir: the directory for caching tokenizers
- hash-seed: seed for hash generation (if not set, is read from PYTHONHASHSEED environment variable)
- zmq-endpoint: ZMQ address to publish events
- zmq-max-connect-attempts: the maximum number of ZMQ connection attempts, defaults to 0, maximum: 10
- event-batch-size: the maximum number of kv-cache events to be sent together, defaults to 16
- failure-injection-rate: probability (0-100) of injecting failures, optional, default is 0
- failure-types: list of specific failure types to inject (rate_limit, invalid_api_key, context_length, server_error, invalid_request, model_not_found), optional, if empty all types are used
- fake-metrics: represents a predefined set of metrics to be sent to Prometheus as a substitute for the real metrics. When specified, only these fake metrics will be reported — real metrics and fake metrics will never be reported together. The set should include values for- running-requests
- waiting-requests
- kv-cache-usage
- loras- an array containing LoRA information objects, each with the fields:- running(a comma-separated list of LoRAs in use by running requests),- waiting(a comma-separated list of LoRAs to be used by waiting requests), and- timestamp(seconds since Jan 1 1970, the timestamp of this metric).
- ttft-buckets-values- array of values for time-to-first-token buckets, each value in this array is a value for the corresponding bucket. Array may contain less values than number of buckets, all trailing missing values assumed as 0. Buckets upper boundaries are: 0.001, 0.005, 0.01, 0.02, 0.04, 0.06, 0.08, 0.1, 0.25, 0.5, 0.75, 1.0, 2.5, 5.0, 7.5, 10.0, 20.0, 40.0, 80.0, 160.0, 640.0, 2560.0, +Inf.
- tpot-buckets-values- array of values for time-per-output-token buckets, each value in this array is a value for the corresponding bucket. Array may contain less values than number of buckets, all trailing missing values assumed as 0. Buckets upper boundaries are: 0.01, 0.025, 0.05, 0.075, 0.1, 0.15, 0.2, 0.3, 0.4, 0.5, 0.75, 1.0, 2.5, 5.0, 7.5, 10.0, 20.0, 40.0, 80.0, +Inf.
 
 Example:
 --fake-metrics '{"running-requests":10,"waiting-requests":30,"kv-cache-usage":0.4,"loras":[{"running":"lora4,lora2","waiting":"lora3","timestamp":1257894567},{"running":"lora4,lora3","waiting":"","timestamp":1257894569}]}'
- data-parallel-size: number of ranks to run in Data Parallel deployment, from 1 to 8, default is 1. The ports will be assigned as follows: rank 0 will run on the configured- port, rank 1 on- port+1, etc.
- dataset-path: Optional local file path to the SQLite database file used for generating responses from a dataset.- If not set, hardcoded preset responses will be used.
- If set but the file does not exist the dataset-urlwill be used to download the database to the path specified bydataset-path.
- If the file exists but is currently occupied by another process, responses will be randomly generated from preset text (the same behavior as if the path were not set).
- Responses are retrieved from the dataset by the hash of the conversation history, with a fallback to a random dataset response, constrained by the maximum output tokens and EoS token handling, if no matching history is found.
- Refer to llm-d converted ShareGPT for detailed information on the expected format of the SQLite database file.
 
- dataset-url: Optional URL for downloading the SQLite database file used for response generation.- This parameter is only used if the dataset-pathis also set and the file does not exist at that path.
- If the file needs to be downloaded, it will be saved to the location specified by dataset-path.
- If the file already exists at the dataset-path, it will not be downloaded again
- Example URL https://huggingface.co/datasets/hf07397/inference-sim-datasets/resolve/91ffa7aafdfd6b3b1af228a517edc1e8f22cd274/huggingface/ShareGPT_Vicuna_unfiltered/conversations.sqlite3
 
- This parameter is only used if the 
- dataset-in-memory: If true, the entire dataset will be loaded into memory for faster access. This may require significant memory depending on the size of the dataset. Default is false.
In addition, as we are using klog, the following parameters are available:
- add_dir_header: if true, adds the file directory to the header of the log messages
- alsologtostderr: log to standard error as well as files (no effect when -logtostderr=true)
- log_backtrace_at: when logging hits line file:N, emit a stack trace (default :0)
- log_dir: if non-empty, write log files in this directory (no effect when -logtostderr=true)
- log_file: if non-empty, use this log file (no effect when -logtostderr=true)
- log_file_max_size: defines the maximum size a log file can grow to (no effect when -logtostderr=true). Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800)
- logtostderr: log to standard error instead of files (default true)
- one_output: if true, only write logs to their native severity level (vs also writing to each lower severity level; no effect when -logtostderr=true)
- skip_headers: if true, avoid header prefixes in the log messages
- skip_log_headers: if true, avoid headers when opening log files (no effect when -logtostderr=true)
- stderrthreshold: logs at or above this threshold go to stderr when writing to files and stderr (no effect when -logtostderr=true or -alsologtostderr=true) (default 2)
- v: number for the log level verbosity
- vmodule: comma-separated list of pattern=N settings for file-filtered logging
- POD_NAME: the simulator pod name. If defined, the response will contain the HTTP header- x-inference-podwith this value, and the HTTP header- x-inference-portwith the port that the request was received on
- POD_NAMESPACE: the simulator pod namespace. If defined, the response will contain the HTTP header- x-inference-namespacewith this value
- max-running-requestswas replaced by- max-num-seqs
- lorawas replaced by- lora-modules, which is now a list of JSON strings, e.g, '{"name": "name", "path": "lora_path", "base_model_name": "id"}'
To build a Docker image of the vLLM Simulator, run:
make image-buildPlease note that the default image tag is ghcr.io/llm-d/llm-d-inference-sim:dev. 
The following environment variables can be used to change the image tag: REGISTRY, SIM_TAG, IMAGE_TAG_BASE or IMG.
Note: On macOS, use make image-build TARGETOS=linux to pull the correct base image.
To run the vLLM Simulator image under Docker, run:
docker run --rm --publish 8000:8000 ghcr.io/llm-d/llm-d-inference-sim:dev  --port 8000 --model "Qwen/Qwen2.5-1.5B-Instruct"  --lora-modules '{"name":"tweet-summary-0"}' '{"name":"tweet-summary-1"}'Note: To run the vLLM Simulator with the latest release version, in the above docker command replace dev with the current release which can be found on GitHub.
Note: The above command exposes the simulator on port 8000, and serves the Qwen/Qwen2.5-1.5B-Instruct model.
To build the vLLM simulator to run locally as an executable, run:
make buildTo run the vLLM simulator in a standalone test environment, run:
./bin/llm-d-inference-sim --model my_model --port 8000To run the vLLM simulator in a Kubernetes cluster, run:
kubectl apply -f manifests/deployment.yamlWhen testing locally with kind, build the docker image with make build-image then load into the cluster:
kind load --name kind docker-image ghcr.io/llm-d/llm-d-inference-sim:devUpdate the deployment.yaml file to use the dev tag.
To verify the deployment is available, run:
kubectl get deployment vllm-llama3-8b-instruct
kubectl get service vllm-llama3-8b-instruct-svcUse kubectl port-forward to expose the service on your local machine:
kubectl port-forward svc/vllm-llama3-8b-instruct-svc 8000:8000Test the API with curl
curl -X POST http://localhost:8000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "meta-llama/Llama-3.1-8B-Instruct",
    "messages": [
      {"role": "user", "content": "Hello!"}
    ]
  }'