@@ -144,27 +144,41 @@ class GenerationConfig:
144
144
Note: The default value varies by model, see the
145
145
`Model.top_k` attribute of the `Model` returned the
146
146
`genai.get_model` function.
147
-
147
+ seed:
148
+ Optional. Seed used in decoding. If not set, the request uses a randomly generated seed.
148
149
response_mime_type:
149
150
Optional. Output response mimetype of the generated candidate text.
150
151
151
152
Supported mimetype:
152
153
`text/plain`: (default) Text output.
154
+ `text/x-enum`: for use with a string-enum in `response_schema`
153
155
`application/json`: JSON response in the candidates.
154
156
155
157
response_schema:
156
158
Optional. Specifies the format of the JSON requested if response_mime_type is
157
159
`application/json`.
160
+ presence_penalty:
161
+ Optional.
162
+ frequency_penalty:
163
+ Optional.
164
+ response_logprobs:
165
+ Optional. If true, export the `logprobs` results in response.
166
+ logprobs:
167
+ Optional. Number of candidates of log probabilities to return at each step of decoding.
158
168
"""
159
-
160
169
candidate_count : int | None = None
161
170
stop_sequences : Iterable [str ] | None = None
162
171
max_output_tokens : int | None = None
163
172
temperature : float | None = None
164
173
top_p : float | None = None
165
174
top_k : int | None = None
175
+ seed : int | None = None
166
176
response_mime_type : str | None = None
167
177
response_schema : protos .Schema | Mapping [str , Any ] | type | None = None
178
+ presence_penalty :float | None = None
179
+ frequency_penalty : float | None = None
180
+ response_logprobs : bool | None = None
181
+ logprobs : int | None = None
168
182
169
183
170
184
GenerationConfigType = Union [protos .GenerationConfig , GenerationConfigDict , GenerationConfig ]
0 commit comments