You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/sphinx_doc/source/tutorial/trinity_configs.md
+4-2Lines changed: 4 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -321,9 +321,10 @@ explorer:
321
321
max_retry_times: 2
322
322
env_vars: {}
323
323
rollout_model:
324
-
engine_type: vllm_async
324
+
engine_type: vllm
325
325
engine_num: 1
326
326
tensor_parallel_size: 1
327
+
enable_history: False
327
328
auxiliary_models:
328
329
- model_path: /PATH/TO/MODEL
329
330
tensor_parallel_size: 1
@@ -336,9 +337,10 @@ explorer:
336
337
- `max_timeout`: Maximum time (in seconds) for a workflow to complete.
337
338
- `max_retry_times`: Maximum number of retries for a workflow.
338
339
- `env_vars`: Environment variables to be set for every workflow runners.
339
-
- `rollout_model.engine_type`: Type of inference engine. Options: `vllm_async` (recommended), `vllm`.
340
+
- `rollout_model.engine_type`: Type of inference engine. For now, only `vllm_async` and `vllm` is supported, they have the same meaning and both use the asynchronous engine. In subsequent versions, only `vllm` may be retained for simplicity.
340
341
- `rollout_model.engine_num`: Number of inference engines.
341
342
- `rollout_model.tensor_parallel_size`: Degree of tensor parallelism.
343
+
- `rollout_model.enable_history`: Whether to enable model call history recording. If set to `True`, the model wrapper automatically records the return experiences of model calls. Please periodically extract the history via `extract_experience_from_history` to avoid out-of-memory issues. Default is `False`.
342
344
- `auxiliary_models`: Additional models used for custom workflows.
343
345
- `eval_interval`: Interval (in steps) for evaluating the model.
344
346
- `eval_on_startup`: Whether to evaluate the model on startup. More precisely, at step 0 with the original model, so it will not be triggered when restarting.
Copy file name to clipboardExpand all lines: docs/sphinx_doc/source/tutorial/trinity_programming_guide.md
+1Lines changed: 1 addition & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -122,6 +122,7 @@ During initialization, `Workflow` receives the following parameters:
122
122
123
123
```{tip}
124
124
You can switch to using the OpenAI API by setting `explorer.rollout_model.enable_openai_api` to `true` in your config file and calling `model.get_openai_client()` to get an `openai.OpenAI` instance in your workflow.
125
+
And the `model` field when calling openai API can be obtained via `openai_client.models.list().data[0].id`.
125
126
```
126
127
127
128
Here's an example of initializing a simple workflow using only `raw_task` and `rollout_args`. In more complex cases, you can use the `format_args` for further customization.
0 commit comments