You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
NOTE: The Groq API is not fully compatible with the OpenAI API.
13
-
Be aware for the following https://console.groq.com/docs/openai[compatability constrains].
13
+
Be aware for the following https://console.groq.com/docs/openai[compatibility constrains].
14
14
Additionally, currently Groq doesn't support multimodal messages.
15
15
16
16
Check the https://github.com/spring-projects/spring-ai/blob/main/models/spring-ai-openai/src/test/java/org/springframework/ai/openai/chat/proxy/GroqWithOpenAiChatModelIT.java[GroqWithOpenAiChatModelIT.java] tests
@@ -24,7 +24,7 @@ The Spring AI project defines a configuration property named `spring.ai.openai.a
24
24
* Set the Groq URL.
25
25
You have to set the `spring.ai.openai.base-url` property to `https://api.groq.com/openai`.
26
26
* Select a https://console.groq.com/docs/models[Groq Model].
27
-
Use the `spring.ai.openai.chat.model=<model name>` property to set the Model.
27
+
Use the `spring.ai.openai.chat.options.model=<model name>` property to set the Model.
28
28
29
29
Exporting an environment variable is one way to set that configuration property:
30
30
@@ -115,7 +115,7 @@ The prefix `spring.ai.openai.chat` is the property prefix that lets you configur
| spring.ai.openai.chat.base-url | Optional overrides the spring.ai.openai.base-url to provide chat specific url. Must be set to `https://api.groq.com/openai` | -
117
117
| spring.ai.openai.chat.api-key | Optional overrides the spring.ai.openai.api-key to provide chat specific api-key | -
118
-
| spring.ai.openai.chat.options.model | The avalable https://console.groq.com/docs/models[model] names are `llama3-8b-8192`, `llama3-70b-8192`, `mixtral-8x7b-32768`, `gemma-7b-it`. | -
118
+
| spring.ai.openai.chat.options.model | The https://console.groq.com/docs/models[available model] names are `llama3-8b-8192`, `llama3-70b-8192`, `mixtral-8x7b-32768`, `gemma2-9b-it`. | -
119
119
| spring.ai.openai.chat.options.temperature | The sampling temperature to use that controls the apparent creativity of generated completions. Higher values will make output more random while lower values will make results more focused and deterministic. It is not recommended to modify temperature and top_p for the same completions request as the interaction of these two settings is difficult to predict. | 0.8
120
120
| spring.ai.openai.chat.options.frequencyPenalty | Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim. | 0.0f
121
121
| spring.ai.openai.chat.options.maxTokens | The maximum number of tokens to generate in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length. | -
0 commit comments