You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat(tongyi): add qwen3-max-2026-01-23 model (#2499)
* feat(tongyi): add qwen3-max-2026-01-23 model
Add new qwen3-max-2026-01-23 model configuration file.
Co-Authored-By: Claude (glm-4.7) <[email protected]>
* feat(tongyi): add qwen3-max-2026-01-23 model
Add new qwen3-max-2026-01-23 model configuration file.
Co-Authored-By: Claude (glm-4.7) <[email protected]>
* feat(tongyi): add qwen3-max-2026-01-23 model
Add new qwen3-max-2026-01-23 model configuration file.
Co-Authored-By: Claude (glm-4.7) <[email protected]>
---------
Co-authored-by: 泠悟 <[email protected]>
Co-authored-by: Claude (glm-4.7) <[email protected]>
en_US: Used to control the degree of randomness and diversity. Specifically, the temperature value controls the degree to which the probability distribution of each candidate word is smoothed when generating text. A higher temperature value will reduce the peak value of the probability distribution, allowing more low-probability words to be selected, and the generated results will be more diverse; while a lower temperature value will enhance the peak value of the probability distribution, making it easier for high-probability words to be selected, the generated results are more certain.
en_US: It is used to specify the maximum number of tokens when the model generates content. It defines the upper limit of generation, but does not guarantee that this number will be generated every time.
en_US: The probability threshold of the kernel sampling method during the generation process. For example, when the value is 0.8, only the smallest set of the most likely tokens with a sum of probabilities greater than or equal to 0.8 is retained as the candidate set. The value range is (0,1.0). The larger the value, the higher the randomness generated; the lower the value, the higher the certainty generated.
en_US: The size of the sample candidate set when generated. For example, when the value is 50, only the 50 highest-scoring tokens in a single generation form a randomly sampled candidate set. The larger the value, the higher the randomness generated; the smaller the value, the higher the certainty generated.
en_US: The random number seed used when generating, the user controls the randomness of the content generated by the model. Supports unsigned 64-bit integers, default value is 1234. When using seed, the model will try its best to generate the same or similar results, but there is currently no guarantee that the results will be exactly the same every time.
en_US: Used to control the repeatability when generating models. Increasing repetition_penalty can reduce the duplication of model generation. 1.0 means no punishment.
en_US: The model has a built-in Internet search service. This parameter controls whether the model refers to Internet search results when generating text. When Internet search is enabled, the model will use the search results as reference information in the text generation process, but the model will "judge" whether to use Internet search results based on its internal logic.
80
+
- name: response_format
81
+
label:
82
+
zh_Hans: 回复格式
83
+
en_US: Response Format
84
+
type: string
85
+
help:
86
+
zh_Hans: 指定模型必须输出的格式
87
+
en_US: specifying the format that the model must output
88
+
required: false
89
+
options:
90
+
- text
91
+
- json_object
92
+
pricing:
93
+
input: '0.006'# 0<Token≤32K
94
+
output: '0.024'# 0<Token≤32K
95
+
# Tiered pricing per 1k tokens: 0<Token≤32K=0.006/0.024
0 commit comments