Skip to content

Commit a7d4290

Browse files
committed
address feedback
1 parent 4477219 commit a7d4290

File tree

1 file changed

+92
-93
lines changed

1 file changed

+92
-93
lines changed

docs/cody/enterprise/model-config-examples.mdx

Lines changed: 92 additions & 93 deletions
Original file line numberDiff line numberDiff line change
@@ -191,57 +191,57 @@ In the configuration above,
191191
```json
192192
"cody.enabled": true,
193193
"modelConfiguration": {
194-
"sourcegraph": null,
195-
"providerOverrides": [
196-
{
197-
"id": "fireworks",
198-
"displayName": "Fireworks",
199-
"serverSideConfig": {
200-
"type": "fireworks",
201-
"accessToken": "token",
202-
"endpoint": "https://api.fireworks.ai/inference/v1/completions"
203-
}
204-
}
205-
],
206-
"modelOverrides": [
207-
{
208-
"modelRef": "fireworks::v1::mixtral-8x22b-instruct",
209-
"displayName": "Mixtral 8x22B",
210-
"modelName": "accounts/fireworks/models/mixtral-8x22b-instruct",
211-
"capabilities": ["chat"],
212-
"category": "other",
213-
"status": "stable",
214-
"contextWindow": {
215-
"maxInputTokens": 7000,
216-
"maxOutputTokens": 4000
217-
}
218-
},
219-
{
220-
"modelRef": "fireworks::v1::starcoder-16b",
221-
"modelName": "accounts/fireworks/models/starcoder-16b",
222-
"displayName": "(Fireworks) Starcoder 16B",
223-
"contextWindow": {
224-
"maxInputTokens": 8192,
225-
"maxOutputTokens": 4096
226-
},
227-
"capabilities": ["autocomplete"],
228-
"category": "balanced",
229-
"status": "stable"
230-
}
231-
],
232-
"defaultModels": {
233-
"chat": "fireworks::v1::mixtral-8x22b-instruct",
234-
"fastChat": "fireworks::v1::mixtral-8x22b-instruct",
235-
"autocomplete": "fireworks::v1::starcoder-16b"
236-
}
194+
"sourcegraph": null,
195+
"providerOverrides": [
196+
{
197+
"id": "fireworks",
198+
"displayName": "Fireworks",
199+
"serverSideConfig": {
200+
"type": "fireworks",
201+
"accessToken": "token",
202+
"endpoint": "https://api.fireworks.ai/inference/v1/completions"
203+
}
204+
}
205+
],
206+
"modelOverrides": [
207+
{
208+
"modelRef": "fireworks::v1::mixtral-8x7b-instruct",
209+
"displayName": "Mixtral 8x7B",
210+
"modelName": "accounts/fireworks/models/mixtral-8x7b-instruct",
211+
"capabilities": ["chat"],
212+
"category": "other",
213+
"status": "stable",
214+
"contextWindow": {
215+
"maxInputTokens": 7000,
216+
"maxOutputTokens": 4000
217+
}
218+
},
219+
{
220+
"modelRef": "fireworks::v1::starcoder-16b",
221+
"modelName": "accounts/fireworks/models/starcoder-16b",
222+
"displayName": "(Fireworks) Starcoder 16B",
223+
"contextWindow": {
224+
"maxInputTokens": 8192,
225+
"maxOutputTokens": 4096
226+
},
227+
"capabilities": ["autocomplete"],
228+
"category": "balanced",
229+
"status": "stable"
230+
}
231+
],
232+
"defaultModels": {
233+
"chat": "fireworks::v1::mixtral-8x7b-instruct",
234+
"fastChat": "fireworks::v1::mixtral-8x7b-instruct",
235+
"autocomplete": "fireworks::v1::starcoder-16b"
236+
}
237237
}
238238
```
239239

240240
In the configuration above,
241241

242242
- Set up a provider override for Fireworks, routing requests for this provider directly to the specified Fireworks endpoint (bypassing Cody Gateway)
243243
- Add two Fireworks models:
244-
- `"fireworks::v1::mixtral-8x22b-instruct"` with "chat" capabiity - used for "chat" and "fastChat"
244+
- `"fireworks::v1::mixtral-8x7b-instruct"` with "chat" capabiity - used for "chat" and "fastChat"
245245
- `"fireworks::v1::starcoder-16b"` with "autocomplete" capability - used for "autocomplete"
246246

247247
</Accordion>
@@ -358,7 +358,6 @@ In the configuration above,
358358
}
359359
```
360360

361-
In the configuration above,
362361
In the configuration above,
363362

364363
- Set up a provider override for Azure OpenAI, routing requests for this provider directly to the specified Azure OpenAI endpoint (bypassing Cody Gateway).
@@ -461,6 +460,51 @@ In the configuration above,
461460

462461
</Accordion>
463462

463+
<Accordion title="Google Gemini">
464+
465+
```json
466+
"modelConfiguration": {
467+
"sourcegraph": null,
468+
"providerOverrides": [
469+
{
470+
"id": "google",
471+
"displayName": "Google Gemini",
472+
"serverSideConfig": {
473+
"type": "google",
474+
"accessToken": "token",
475+
"endpoint": "https://generativelanguage.googleapis.com/v1beta/models"
476+
}
477+
}
478+
],
479+
"modelOverrides": [
480+
{
481+
"modelRef": "google::v1::gemini-1.5-pro",
482+
"displayName": "Gemini 1.5 Pro",
483+
"modelName": "gemini-1.5-pro",
484+
"capabilities": ["chat", "autocomplete"],
485+
"category": "balanced",
486+
"status": "stable",
487+
"contextWindow": {
488+
"maxInputTokens": 45000,
489+
"maxOutputTokens": 4000
490+
}
491+
}
492+
],
493+
"defaultModels": {
494+
"chat": "google::v1::gemini-1.5-pro",
495+
"fastChat": "google::v1::gemini-1.5-pro",
496+
"autocomplete": "google::v1::gemini-1.5-pro"
497+
}
498+
}
499+
```
500+
501+
In the configuration above,
502+
503+
- Set up a provider override for Google Gemini, routing requests for this provider directly to the specified endpoint (bypassing Cody Gateway)
504+
- Add the `"google::v1::gemini-1.5-pro"` model, which is used for all Cody features. We do not add other models for simplicity, as adding multiple models is already covered in the examples above
505+
506+
</Accordion>
507+
464508
<Accordion title="Google Vertex (Anthropic)">
465509

466510
```json
@@ -520,51 +564,6 @@ In the configuration above,
520564

521565
</Accordion>
522566

523-
<Accordion title="Google Gemini">
524-
525-
```json
526-
"modelConfiguration": {
527-
"sourcegraph": null,
528-
"providerOverrides": [
529-
{
530-
"id": "google",
531-
"displayName": "Google Gemini",
532-
"serverSideConfig": {
533-
"type": "google",
534-
"accessToken": "token",
535-
"endpoint": "https://generativelanguage.googleapis.com/v1beta/models"
536-
}
537-
}
538-
],
539-
"modelOverrides": [
540-
{
541-
"modelRef": "google::v1::gemini-1.5-pro",
542-
"displayName": "Gemini 1.5 Pro",
543-
"modelName": "gemini-1.5-pro",
544-
"capabilities": ["chat", "autocomplete"],
545-
"category": "balanced",
546-
"status": "stable",
547-
"contextWindow": {
548-
"maxInputTokens": 45000,
549-
"maxOutputTokens": 4000
550-
}
551-
}
552-
],
553-
"defaultModels": {
554-
"chat": "google::v1::gemini-1.5-pro",
555-
"fastChat": "google::v1::gemini-1.5-pro",
556-
"autocomplete": "google::v1::gemini-1.5-pro"
557-
}
558-
}
559-
```
560-
561-
In the configuration above,
562-
563-
- Set up a provider override for Google Gemini, routing requests for this provider directly to the specified endpoint (bypassing Cody Gateway)
564-
- Add the `"google::v1::gemini-1.5-pro"` model, which is used for all Cody features. We do not add other models for simplicity, as adding multiple models is already covered in the examples above
565-
566-
</Accordion>
567-
568567
<Accordion title="AWS Bedrock">
569568

570569
```json
@@ -667,12 +666,12 @@ Example configuration:
667666
},
668667
{
669668
"provider": "fireworks",
670-
"model": "mixtral-8x22b-instruct@v1",
669+
"model": "mixtral-8x7b-instruct@v1",
671670
"override": {
672-
"displayName": "(Fireworks) Mixtral 8x22b Instruct",
671+
"displayName": "(Fireworks) Mixtral 8x7b Instruct",
673672
"serverSideConfig": {
674673
"type": "openaicompatible",
675-
"apiModel": "accounts/fireworks/models/mixtral-8x22b-instruct"
674+
"apiModel": "accounts/fireworks/models/mixtral-8x7b-instruct"
676675
}
677676
}
678677
},

0 commit comments

Comments
 (0)