diff --git a/docs/api/google/generativeai/chat_async.md b/docs/api/google/generativeai/chat_async.md index 614456c1f..c7e79aadd 100644 --- a/docs/api/google/generativeai/chat_async.md +++ b/docs/api/google/generativeai/chat_async.md @@ -1,67 +1,27 @@ -description: Calls the API to initiate a chat with a model using provided parameters -
- - -
- -# google.generativeai.chat_async - - - - - - - - - -Calls the API to initiate a chat with a model using provided parameters - - - - - - - - - - - + 
+ `model` + + Which model to call, as a string or a types.Model. +
+ `context` + + Text that should be provided to the model first, to ground the response. If not empty, this `context` will be given to the model first before the @@ -77,12 +37,16 @@ Examples: Anything included in this field will take precedence over history in `messages` if the total input size exceeds the model's Model.input_token_limit. +
+ `examples` + + Examples of what the model should generate. This includes both the user input and the response that the model should @@ -92,34 +56,46 @@ These `examples` are treated identically to conversation messages except that they take precedence over the history in `messages`: If the total input size exceeds the model's `input_token_limit` the input will be truncated. Items will be dropped from `messages` before `examples` +
+ `messages` + + A snapshot of the conversation history sorted chronologically. Turns alternate between two authors. If the total input size exceeds the model's `input_token_limit` the input will be truncated: The oldest items will be dropped from `messages`. +
+ `temperature` + + Controls the randomness of the output. Must be positive. Typical values are in the range: `[0.0,1.0]`. Higher values produce a more random and varied response. A temperature of zero will be deterministic. +
+ `candidate_count` + + The **maximum** number of generated response messages to return. This value must be between `[1, 8]`, inclusive. If unset, this @@ -128,22 +104,30 @@ will default to `1`. Note: Only unique candidates are returned. Higher temperatures are more likely to produce unique candidates. Setting `temperature=0.0` will always return 1 candidate regardless of the `candidate_count`. +
+ `top_k` + + The API uses combined [nucleus](https://arxiv.org/abs/1904.09751) and top-k sampling. `top_k` sets the maximum number of tokens to sample from on each step. +
+ `top_p` + + The API uses combined [nucleus](https://arxiv.org/abs/1904.09751) and top-k sampling. @@ -155,29 +139,42 @@ top-k sampling. as `[0.625, 0.25, 0.125, 0, 0, 0]`. Typical values are in the `[0.9, 1.0]` range. +
+ `prompt` + + You may pass a types.MessagePromptOptions **instead** of a setting `context`/`examples`/`messages`, but not both. +
+ `client` + + If you're not relying on the default client, you pass a `glm.DiscussServiceClient` instead. +
+ `request_options` + + Options for the request. +