You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A new unit test has been added to explicitly verify that headers can be removed by setting their value to `None`. Another unit test has been added to ensure that this functionality works correctly when using the CLI.
The documentation for the `--backend-args` CLI option has been updated to explicitly mention that headers can be removed by setting their value to `null`.
|`--scenario <PATH or NAME>`| The name of a builtin scenario or path to a scenario configuration file. Options specified on the command line will override the scenario file. |
14
14
15
15
### Target and Backend Configuration
16
16
17
17
These options configure how `guidellm` connects to the system under test.
18
18
19
-
| Option | Description |
20
-
| --- | --- |
21
-
|`--target <URL>`|**Required.** The endpoint of the target system, e.g., `http://localhost:8080`. Can also be set with the `GUIDELLM__OPENAI__BASE_URL` environment variable. |
22
-
|`--backend-type <TYPE>`| The type of backend to use. Defaults to `openai_http`. |
|`--target <URL>`|**Required.** The endpoint of the target system, e.g., `http://localhost:8080`. Can also be set with the `GUIDELLM__OPENAI__BASE_URL` environment variable.|
22
+
|`--backend-type <TYPE>`| The type of backend to use. Defaults to `openai_http`. |
23
23
|`--backend-args <JSON>`| A JSON string for backend-specific arguments. For example: `--backend-args '{"headers": {"Authorization": "Bearer my-token"}, "verify": false}'` to pass custom headers and disable certificate verification. |
24
-
|`--model <NAME>`| The ID of the model to benchmark within the backend. |
24
+
|`--model <NAME>`| The ID of the model to benchmark within the backend.|
25
25
26
26
### Data and Request Configuration
27
27
28
28
These options define the data to be used for benchmarking and how requests will be generated.
29
29
30
-
| Option | Description |
31
-
| --- | --- |
32
-
|`--data <SOURCE>`| The data source. This can be a HuggingFace dataset ID, a path to a local data file, or a synthetic data configuration. See the [Data Formats Guide](./data_formats.md) for more details. |
33
-
|`--rate-type <TYPE>`| The type of request generation strategy to use (e.g., `constant`, `poisson`, `sweep`). |
34
-
|`--rate <NUMBER>`| The rate of requests per second for `constant` or `poisson` strategies, or the number of steps for a `sweep`. |
35
-
|`--max-requests <NUMBER>`| The maximum number of requests to run for each benchmark. |
36
-
|`--max-seconds <NUMBER>`| The maximum number of seconds to run each benchmark for. |
|`--data <SOURCE>`| The data source. This can be a HuggingFace dataset ID, a path to a local data file, or a synthetic data configuration. See the [Data Formats Guide](./data_formats.md) for more details. |
33
+
|`--rate-type <TYPE>`| The type of request generation strategy to use (e.g., `constant`, `poisson`, `sweep`).|
34
+
|`--rate <NUMBER>`| The rate of requests per second for `constant` or `poisson` strategies, or the number of steps for a `sweep`.|
35
+
|`--max-requests <NUMBER>`| The maximum number of requests to run for each benchmark. |
36
+
|`--max-seconds <NUMBER>`| The maximum number of seconds to run each benchmark for.|
Copy file name to clipboardExpand all lines: docs/guides/configuration.md
+20-19Lines changed: 20 additions & 19 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,16 +5,18 @@ The `guidellm` application can be configured using command-line arguments, envir
5
5
## Configuration Methods
6
6
7
7
Settings are loaded with the following priority (highest priority first):
8
-
1. Command-line arguments.
9
-
2. Environment variables.
10
-
3. Values in a `.env` file in the directory where the command is run.
11
-
4. Default values.
8
+
9
+
1. Command-line arguments.
10
+
2. Environment variables.
11
+
3. Values in a `.env` file in the directory where the command is run.
12
+
4. Default values.
12
13
13
14
## Environment Variable Format
14
15
15
16
All settings can be configured using environment variables. The variables must be prefixed with `GUIDELLM__`, and nested settings are separated by a double underscore `__`.
16
17
17
18
For example, to set the `api_key` for the `openai` backend, you would use the following environment variable:
You can configure the connection to the target system using environment variables. This is an alternative to using the `--target-*` command-line flags.
25
27
26
-
| Environment Variable | Description | Example |
27
-
| --- | --- | --- |
28
-
|`GUIDELLM__OPENAI__BASE_URL`| The endpoint of the target system. Equivalent to the `--target` CLI option. |`export GUIDELLM__OPENAI__BASE_URL="http://localhost:8080"`|
29
-
|`GUIDELLM__OPENAI__API_KEY`| The API key to use for bearer token authentication. |`export GUIDELLM__OPENAI__API_KEY="your-secret-api-key"`|
30
-
|`GUIDELLM__OPENAI__BEARER_TOKEN`| The full bearer token to use for authentication. |`export GUIDELLM__OPENAI__BEARER_TOKEN="Bearer your-secret-token"`|
31
-
|`GUIDELLM__OPENAI__HEADERS`| A JSON string representing a dictionary of headers to send to the target. These headers will override any default headers. |`export GUIDELLM__OPENAI__HEADERS='{"Authorization": "Bearer my-token"}'`|
32
-
|`GUIDELLM__OPENAI__ORGANIZATION`| The OpenAI organization to use for requests. |`export GUIDELLM__OPENAI__ORGANIZATION="org-12345"`|
33
-
|`GUIDELLM__OPENAI__PROJECT`| The OpenAI project to use for requests. |`export GUIDELLM__OPENAI__PROJECT="proj-67890"`|
34
-
|`GUIDELLM__OPENAI__VERIFY`| Set to `false` or `0` to disable certificate verification. |`export GUIDELLM__OPENAI__VERIFY=false`|
35
-
|`GUIDELLM__OPENAI__MAX_OUTPUT_TOKENS`| The default maximum number of tokens to request for completions. |`export GUIDELLM__OPENAI__MAX_OUTPUT_TOKENS=2048`|
|`GUIDELLM__OPENAI__BASE_URL`| The endpoint of the target system. Equivalent to the `--target` CLI option. |`export GUIDELLM__OPENAI__BASE_URL="http://localhost:8080"`|
31
+
|`GUIDELLM__OPENAI__API_KEY`| The API key to use for bearer token authentication. |`export GUIDELLM__OPENAI__API_KEY="your-secret-api-key"`|
32
+
|`GUIDELLM__OPENAI__BEARER_TOKEN`| The full bearer token to use for authentication. |`export GUIDELLM__OPENAI__BEARER_TOKEN="Bearer your-secret-token"`|
33
+
|`GUIDELLM__OPENAI__HEADERS`| A JSON string representing a dictionary of headers to send to the target. These headers will override any default headers. |`export GUIDELLM__OPENAI__HEADERS='{"Authorization": "Bearer my-token"}'`|
34
+
|`GUIDELLM__OPENAI__ORGANIZATION`| The OpenAI organization to use for requests. |`export GUIDELLM__OPENAI__ORGANIZATION="org-12345"`|
35
+
|`GUIDELLM__OPENAI__PROJECT`| The OpenAI project to use for requests. |`export GUIDELLM__OPENAI__PROJECT="proj-67890"`|
36
+
|`GUIDELLM__OPENAI__VERIFY`| Set to `false` or `0` to disable certificate verification. |`export GUIDELLM__OPENAI__VERIFY=false`|
37
+
|`GUIDELLM__OPENAI__MAX_OUTPUT_TOKENS`| The default maximum number of tokens to request for completions. |`export GUIDELLM__OPENAI__MAX_OUTPUT_TOKENS=2048`|
36
38
37
39
### General HTTP Settings
38
40
39
41
These settings control the behavior of the underlying HTTP client.
40
42
41
-
| Environment Variable | Description |
42
-
| --- | --- |
43
-
|`GUIDELLM__REQUEST_TIMEOUT`| The timeout in seconds for HTTP requests. Defaults to 300. |
44
-
|`GUIDELLM__REQUEST_HTTP2`| Set to `true` or `1` to enable HTTP/2 support. Defaults to true. |
0 commit comments