You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- For lists of items, all elements must be of the same type.
222
222
- A processor/tokenizer is only required if `GUIDELLM__PREFERRED_PROMPT_TOKENS_SOURCE="local"` or `GUIDELLM__PREFERRED_OUTPUT_TOKENS_SOURCE="local"` is set in the environment. In this case, the processor/tokenizer must be specified using the `--processor` argument. If not set, the processor/tokenizer will be set to the model passed in or retrieved from the server.
223
223
224
-
225
224
### ShareGPT Datasets
226
225
227
226
You can use ShareGPT_V3_unfiltered_cleaned_split.json as benchmark datasets.
228
227
229
-
1. Download and prepare the ShareGPT dataset
230
-
You can specify the proportion of data to process by providing a number between 0 and 1 as an argument to the script.
228
+
#### Example Commands
231
229
232
-
```bash
233
-
cd src/guidellm/utils
234
-
pip install -r requirements.txt
235
-
bash prepare_sharegpt_data.sh 1
236
-
```
230
+
Download and prepare the ShareGPT dataset; You can specify the proportion of data to process by providing a number between 0 and 1 as an argument to the script.
237
231
238
-
In this example, 1 indicates processing 100% of the dataset. You can adjust this value as needed.
0 commit comments