You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: Model/Anthropic/models.json
+38-10Lines changed: 38 additions & 10 deletions
Original file line number
Diff line number
Diff line change
@@ -2,31 +2,59 @@
2
2
"chat": {
3
3
"claude-3-opus-20240229": {
4
4
"description": "Most powerful model for highly complex tasks. Top-level performance, intelligence, fluency, and understanding.",
5
-
"inputTokens": 150000,
6
-
"inputTokenPrice": 15.00,
5
+
"inputTokens": 200000,
6
+
"inputTokenPrice": 15,
7
7
"outputTokens": 4096,
8
-
"outputTokenPrice": 75.00
8
+
"outputTokenPrice": 75
9
9
},
10
10
"claude-3-5-sonnet-20240620": {
11
11
"description": "Most intelligent model",
12
12
"inputTokens": 200000,
13
-
"inputTokenPrice": 3.00,
13
+
"inputTokenPrice": 3,
14
14
"outputTokens": 4096,
15
-
"outputTokenPrice": 15.00
15
+
"outputTokenPrice": 15
16
16
},
17
17
"claude-3-sonnet-20240229": {
18
18
"description": "Ideal balance of intelligence and speed for enterprise workloads. Maximum utility at a lower price, dependable, balanced for scaled deployments.",
19
-
"inputTokens": 150000,
20
-
"inputTokenPrice": 3.00,
19
+
"inputTokens": 200000,
20
+
"inputTokenPrice": 3,
21
21
"outputTokens": 4096,
22
-
"outputTokenPrice": 15.00
22
+
"outputTokenPrice": 15
23
23
},
24
24
"claude-3-haiku-20240307": {
25
25
"description": "Fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance.",
Copy file name to clipboardExpand all lines: Model/Groq/models.json
+12-12Lines changed: 12 additions & 12 deletions
Original file line number
Diff line number
Diff line change
@@ -3,37 +3,37 @@
3
3
"llama3-8b-8192": {
4
4
"description": "LLaMA3 8b",
5
5
"inputTokens": 8192,
6
-
"inputTokenPrice": 0,
6
+
"inputTokenPrice": 0.05,
7
7
"outputTokens": 8192,
8
-
"outputTokenPrice": 0
8
+
"outputTokenPrice": 0.08
9
9
},
10
10
"llama3-70b-8192": {
11
11
"description": "LLaMA3 70b",
12
12
"inputTokens": 8192,
13
-
"inputTokenPrice": 0,
13
+
"inputTokenPrice": 0.59,
14
14
"outputTokens": 8192,
15
-
"outputTokenPrice": 0
15
+
"outputTokenPrice": 0.79
16
16
},
17
17
"llama2-70b-4096": {
18
18
"description": "LLaMA2 70b",
19
19
"inputTokens": 4096,
20
-
"inputTokenPrice": 0,
20
+
"inputTokenPrice": 0.7,
21
21
"outputTokens": 4096,
22
-
"outputTokenPrice": 0
22
+
"outputTokenPrice": 0.8
23
23
},
24
24
"mixtral-8x7b-32768": {
25
25
"description": "Mixtral 8x7b",
26
26
"inputTokens": 32768,
27
-
"inputTokenPrice": 0,
27
+
"inputTokenPrice": 0.24,
28
28
"outputTokens": 32768,
29
-
"outputTokenPrice": 0
29
+
"outputTokenPrice": 0.24
30
30
},
31
31
"gemma-7b-it": {
32
32
"description": "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as a laptop, desktop or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.",
Copy file name to clipboardExpand all lines: Model/Mistral/models.json
+66-17Lines changed: 66 additions & 17 deletions
Original file line number
Diff line number
Diff line change
@@ -4,51 +4,100 @@
4
4
"description": "Our very first. A 7B transformer model, fast-deployed and easily customisable. Small, yet very powerful for a variety of use cases. English and code.",
5
5
"inputTokens": 32000,
6
6
"inputTokenPrice": 0.25,
7
-
"outputTokens": 4096,
7
+
"outputTokens": 8191,
8
8
"outputTokenPrice": 0.25
9
9
},
10
10
"open-mixtral-8x7b": {
11
11
"description": "A 7B sparse Mixture-of-Experts (SMoE). Uses 12B active parameters out of 45B total. Fluent in English, French, Italian, German, Spanish, and strong in code.",
12
12
"inputTokens": 32000,
13
13
"inputTokenPrice": 0.7,
14
-
"outputTokens": 4096,
14
+
"outputTokens": 8191,
15
15
"outputTokenPrice": 0.7
16
16
},
17
17
"open-mixtral-8x22b": {
18
18
"description": "A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B. Fluent in English, French, Italian, German, Spanish, and strong in code.",
19
19
"inputTokens": 64000,
20
-
"inputTokenPrice": 2.0,
21
-
"outputTokens": 4096,
22
-
"outputTokenPrice": 6.0
20
+
"inputTokenPrice": 2,
21
+
"outputTokens": 8191,
22
+
"outputTokenPrice": 6
23
23
},
24
24
"mistral-small-latest": {
25
25
"description": "Cost-efficient reasoning for low-latency workloads. Fluent in English, French, Italian, German, Spanish, and strong in code.",
26
26
"inputTokens": 32000,
27
-
"inputTokenPrice": 2.00,
28
-
"outputTokens": 4096,
29
-
"outputTokenPrice": 6.00
27
+
"inputTokenPrice": 1,
28
+
"outputTokens": 8191,
29
+
"outputTokenPrice": 3
30
30
},
31
31
"mistral-medium-latest": {
32
32
"description": "Balanced reasoning for a wide range of tasks. Fluent in English, French, Italian, German, Spanish, and strong in code.",
33
33
"inputTokens": 32000,
34
-
"inputTokenPrice": 2.70,
35
-
"outputTokens": 4096,
36
-
"outputTokenPrice": 8.10
34
+
"inputTokenPrice": 2.7,
35
+
"outputTokens": 8191,
36
+
"outputTokenPrice": 8.1
37
37
},
38
38
"mistral-large-latest": {
39
39
"description": "Top-tier reasoning for high-complexity tasks. Fluent in English, French, Italian, German, Spanish, and strong in code.",
Copy file name to clipboardExpand all lines: Model/OpenAI/models.json
+41-14Lines changed: 41 additions & 14 deletions
Original file line number
Diff line number
Diff line change
@@ -3,50 +3,77 @@
3
3
"gpt-3.5-turbo": {
4
4
"description": "GPT-3.5 Turbo is a variant of GPT-3.5 that is optimized for speed and can generate responses in a fraction of the time.",
5
5
"inputTokens": 16385,
6
-
"inputTokenPrice": 0.50,
6
+
"inputTokenPrice": 1.5,
7
7
"outputTokens": 4096,
8
-
"outputTokenPrice": 1.50
8
+
"outputTokenPrice": 2
9
9
},
10
10
"gpt-4": {
11
11
"description": "GPT-4 is the latest version of OpenAI's Generative Pre-trained Transformer (GPT) language model. It is capable of generating human-like text and can be used for a wide range of natural language processing tasks.",
12
-
"inputTokens": 81924,
13
-
"inputTokenPrice": 30.00,
12
+
"inputTokens": 8192,
13
+
"inputTokenPrice": 30,
14
14
"outputTokens": 4096,
15
-
"outputTokenPrice": 60.00
15
+
"outputTokenPrice": 60
16
16
},
17
17
"gpt-4-turbo": {
18
18
"description": "GPT-4 Turbo is a variant of GPT-4 that is optimized for speed and can generate responses in a fraction of the time.",
19
19
"inputTokens": 128000,
20
-
"inputTokenPrice": 10.00,
20
+
"inputTokenPrice": 10,
21
21
"outputTokens": 4096,
22
-
"outputTokenPrice": 30.00
22
+
"outputTokenPrice": 30
23
23
},
24
24
"gpt-4o": {
25
25
"description": "GPT-4o has the same high intelligence as GPT-4 Turbo but is much more efficient",
0 commit comments