|
2 | 2 |
|
3 | 3 | public enum CompletionModel |
4 | 4 | { |
5 | | - /* ================================ Completion ================================ */ |
6 | | - TEXT_DAVINCI_003("text-davinci-003", |
7 | | - "Most capable model in the GPT-3 series. Can perform any task the other GPT-3 models can, often with higher quality, longer output and better instruction-following. It can process up to 4,000 tokens per request.", |
8 | | - "Complex intent, cause and effect, creative generation, search, summarization for audience", |
9 | | - 4096), |
10 | | - TEXT_CURIE_001("text-curie-001", |
11 | | - "Very capable, but faster and lower cost than text-davinci-003.", |
12 | | - "Language translation, complex classification, sentiment, summarization", |
| 5 | + /* ================================ Text Model ================================ */ |
| 6 | + ADA("ada", |
| 7 | + "Capable of very simple tasks, usually the fastest model in the GPT-3 series, and lowest cost.", |
| 8 | + null, |
13 | 9 | 2049), |
14 | | - TEXT_BABBAGE_001("text-babbage-001", |
| 10 | + ADA_CODE_SEARCH_CODE("ada-code-search-code", null, null, 4096), |
| 11 | + ADA_CODE_SEARCH_TEXT("ada-code-search-text", null, null, 4096), |
| 12 | + ADA_SEARCH_DOCUMENT("ada-search-document", null, null, 4096), |
| 13 | + ADA_SEARCH_QUERY("ada-search-query", null, null, 4096), |
| 14 | + ADA_SIMILARITY("ada-similarity", null, null, 4096), |
| 15 | + BABBAGE("babbage", |
15 | 16 | "Capable of straightforward tasks, very fast, and lower cost.", |
16 | | - "Moderate classification, semantic search", |
| 17 | + null, |
17 | 18 | 2049), |
| 19 | + BABBAGE_CODE_SEARCH_CODE("babbage-code-search-code", null, null, 4096), |
| 20 | + BABBAGE_CODE_SEARCH_TEXT("babbage-code-search-text", null, null, 4096), |
| 21 | + BABBAGE_SEARCH_DOCUMENT("babbage-search-document", null, null, 4096), |
| 22 | + BABBAGE_SEARCH_QUERY("babbage-search-query", null, null, 4096), |
| 23 | + BABBAGE_SIMILARITY("babbage-similarity", null, null, 4096), |
| 24 | + CODE_DAVINCI_EDIT_001("code-davinci-edit-001", null, null, 4096), |
| 25 | + CODE_SEARCH_ADA_CODE_001("code-search-ada-code-001", null, null, 4096), |
| 26 | + CODE_SEARCH_ADA_TEXT_001("code-search-ada-text-001", null, null, 4096), |
| 27 | + CODE_SEARCH_BABBAGE_CODE_001("code-search-babbage-code-001", null, null, 4096), |
| 28 | + CODE_SEARCH_BABBAGE_TEXT_001("code-search-babbage-text-001", null, null, 4096), |
| 29 | + CURIE("curie", |
| 30 | + "Very capable, but faster and lower cost than Davinci.", |
| 31 | + null, |
| 32 | + 2049), |
| 33 | + CURIE_INSTRUCT_BETA("curie-instruct-beta", null, null, 4096), |
| 34 | + CURIE_SEARCH_DOCUMENT("curie-search-document", null, null, 4096), |
| 35 | + CURIE_SEARCH_QUERY("curie-search-query", null, null, 4096), |
| 36 | + CURIE_SIMILARITY("curie-similarity", null, null, 4096), |
| 37 | + DAVINCI("davinci", |
| 38 | + "Most capable GPT-3 model. Can do any task the other models can do, often with higher quality.", |
| 39 | + null, |
| 40 | + 2049), |
| 41 | + DAVINCI_INSTRUCT_BETA("davinci-instruct-beta", null, null, 4096), |
| 42 | + DAVINCI_SEARCH_DOCUMENT("davinci-search-document", null, null, 4096), |
| 43 | + DAVINCI_SEARCH_QUERY("davinci-search-query", null, null, 4096), |
| 44 | + DAVINCI_SIMILARITY("davinci-similarity", null, null, 4096), |
18 | 45 | TEXT_ADA_001("text-ada-001", |
19 | 46 | "Capable of simple tasks, usually the fastest model in the GPT-3 series, and lowest cost.", |
20 | 47 | "Parsing text, simple classification, address correction, keywords", |
21 | 48 | 2049), |
| 49 | + TEXT_BABBAGE_001("text-babbage-001", |
| 50 | + "Capable of straightforward tasks, very fast, and lower cost.", |
| 51 | + "Moderate classification, semantic search", |
| 52 | + 2049), |
| 53 | + TEXT_CURIE_001("text-curie-001", |
| 54 | + "Very capable, but faster and lower cost than text-davinci-003.", |
| 55 | + "Language translation, complex classification, sentiment, summarization", |
| 56 | + 2049), |
| 57 | + TEXT_DAVINCI_001("text-davinci-001", |
| 58 | + "Most capable model in the GPT-3 series. Can perform any task the other GPT-3 models can, often with higher quality, longer output and better instruction-following. It can process up to 4,000 tokens per request.", |
| 59 | + "Complex intent, cause and effect, creative generation, search, summarization for audience", |
| 60 | + 4096), |
22 | 61 | TEXT_DAVINCI_002("text-davinci-002", |
23 | 62 | "Most capable model in the GPT-3 series. Can perform any task the other GPT-3 models can, often with higher quality, longer output and better instruction-following. It can process up to 4,000 tokens per request.", |
24 | 63 | "Complex intent, cause and effect, creative generation, search, summarization for audience", |
25 | 64 | 4097), |
| 65 | + TEXT_DAVINCI_003("text-davinci-003", |
| 66 | + "Most capable model in the GPT-3 series. Can perform any task the other GPT-3 models can, often with higher quality, longer output and better instruction-following. It can process up to 4,000 tokens per request.", |
| 67 | + "Complex intent, cause and effect, creative generation, search, summarization for audience", |
| 68 | + 4096), |
| 69 | + TEXT_DAVINCI_EDIT_001("text-davinci-edit-001", null, null, 4096), |
| 70 | + TEXT_EMBEDDING_ADA_002("text-embedding-ada-002", null, null, 4096), |
| 71 | + TEXT_SEARCH_ADA_DOC_001("text-search-ada-doc-001", null, null, 4096), |
| 72 | + TEXT_SEARCH_ADA_QUERY_001("text-search-ada-query-001", null, null, 4096), |
| 73 | + TEXT_SEARCH_BABBAGE_DOC_001("text-search-babbage-doc-001", null, null, 4096), |
| 74 | + TEXT_SEARCH_BABBAGE_QUERY_001("text-search-babbage-query-001", null, null, 4096), |
| 75 | + TEXT_SEARCH_CURIE_DOC_001("text-search-curie-doc-001", null, null, 4096), |
| 76 | + TEXT_SEARCH_CURIE_QUERY_001("text-search-curie-query-001", null, null, 4096), |
| 77 | + TEXT_SEARCH_DAVINCI_DOC_001("text-search-davinci-doc-001", null, null, 4096), |
| 78 | + TEXT_SEARCH_DAVINCI_QUERY_001("text-search-davinci-query-001", null, null, 4096), |
| 79 | + TEXT_SIMILARITY_ADA_001("text-similarity-ada-001", null, null, 4096), |
| 80 | + TEXT_SIMILARITY_BABBAGE_001("text-similarity-babbage-001", null, null, 4096), |
| 81 | + TEXT_SIMILARITY_CURIE_001("text-similarity-curie-001", null, null, 4096), |
| 82 | + TEXT_SIMILARITY_DAVINCI_001("text-similarity-davinci-001", null, null, 4096), |
| 83 | + |
| 84 | + /* ================================ Chat Model ================================ */ |
| 85 | + GPT_35_TURBO("gpt-3.5-turbo", |
| 86 | + "Most capable GPT-3.5 model and optimized for chat at 1/10th the cost of text-davinci-003. Will be updated with our latest model iteration 2 weeks after it is released.", |
| 87 | + null, |
| 88 | + 4096), |
| 89 | + |
| 90 | + GPT_35_TURBO_0301("gpt-3.5-turbo-0301", |
| 91 | + null, |
| 92 | + null, |
| 93 | + 4096), |
| 94 | + |
| 95 | + GPT_35_TURBO_0613("gpt-3.5-turbo-0613", |
| 96 | + "Snapshot of gpt-3.5-turbo from June 13th 2023 with function calling data. Unlike gpt-3.5-turbo, this model will not receive updates, and will be deprecated 3 months after a new version is released.", |
| 97 | + null, |
| 98 | + 4096), |
| 99 | + |
| 100 | + GPT_35_TURBO_16K("gpt-3.5-turbo-16k", |
| 101 | + "Same capabilities as the standard gpt-3.5-turbo model but with 4 times the context.", |
| 102 | + null, |
| 103 | + 16384), |
| 104 | + |
| 105 | + GPT_35_TURBO_16K_0613("gpt-3.5-turbo-16k-0613", |
| 106 | + "Snapshot of gpt-3.5-turbo-16k from June 13th 2023. Unlike gpt-3.5-turbo-16k, this model will not receive updates, and will be deprecated 3 months after a new version is released.", |
| 107 | + null, |
| 108 | + 16384), |
26 | 109 |
|
27 | | - /* ================================ Chat Completion ================================ */ |
28 | 110 | GPT_4("gpt-4", |
29 | 111 | "More capable than any GPT-3.5 model, able to do more complex tasks, and optimized for chat. Will be updated with our latest model iteration 2 weeks after it is released.", |
30 | 112 | null, |
31 | 113 | 8192), |
| 114 | + |
| 115 | + GPT_4_0314("gpt-4-0314", |
| 116 | + null, |
| 117 | + null, |
| 118 | + 8192), |
| 119 | + |
32 | 120 | GPT_4_0613("gpt-4-0613", |
33 | 121 | "Snapshot of gpt-4 from June 13th 2023 with function calling data. Unlike gpt-4, this model will not receive updates, and will be deprecated 3 months after a new version is released.", |
34 | 122 | null, |
35 | 123 | 8192), |
| 124 | + |
36 | 125 | GPT_4_32K("gpt-4-32k", |
37 | 126 | "Same capabilities as the base gpt-4 mode but with 4x the context length. Will be updated with our latest model iteration.", |
38 | 127 | null, |
39 | 128 | 32768), |
| 129 | + |
40 | 130 | GPT_4_32K_0613("gpt-4-32k-0613", |
41 | 131 | "Snapshot of gpt-4-32 from June 13th 2023. Unlike gpt-4-32k, this model will not receive updates, and will be deprecated 3 months after a new version is released.", |
42 | 132 | null, |
43 | 133 | 32768), |
44 | | - GPT_35_TURBO("gpt-3.5-turbo", |
45 | | - "Most capable GPT-3.5 model and optimized for chat at 1/10th the cost of text-davinci-003. Will be updated with our latest model iteration 2 weeks after it is released.", |
46 | | - null, |
47 | | - 4096), |
48 | | - GPT_35_TURBO_16K("gpt-3.5-turbo-16k", |
49 | | - "Same capabilities as the standard gpt-3.5-turbo model but with 4 times the context.", |
50 | | - null, |
51 | | - 16384), |
52 | | - GPT_35_TURBO_0613("gpt-3.5-turbo-0613", |
53 | | - "Snapshot of gpt-3.5-turbo from June 13th 2023 with function calling data. Unlike gpt-3.5-turbo, this model will not receive updates, and will be deprecated 3 months after a new version is released.", |
54 | | - null, |
55 | | - 4096), |
56 | | - GPT_35_TURBO_16K_0613("gpt-3.5-turbo-16k-0613", |
57 | | - "Snapshot of gpt-3.5-turbo-16k from June 13th 2023. Unlike gpt-3.5-turbo-16k, this model will not receive updates, and will be deprecated 3 months after a new version is released.", |
58 | | - null, |
59 | | - 16384), |
| 134 | + |
60 | 135 | CODE_DAVINCI_002("code-davinci-002", |
61 | 136 | "Optimized for code-completion tasks", |
62 | 137 | null, |
63 | 138 | 8001), |
64 | 139 |
|
| 140 | + /* ================================ Moderation Model ================================ */ |
65 | 141 | TEXT_MODERATION_LATEST("text-moderation-latest", |
66 | 142 | "Most capable moderation model. Accuracy will be slighlty higher than the stable model.", |
67 | 143 | null, |
68 | 144 | Integer.MAX_VALUE), |
| 145 | + |
69 | 146 | TEXT_MODERATION_STABLE("text-moderation-stable", |
70 | 147 | "Almost as capable as the latest model, but slightly older.\n", |
71 | 148 | null, |
72 | 149 | Integer.MAX_VALUE), |
73 | | - DAVINCI("davinci", |
74 | | - "Most capable GPT-3 model. Can do any task the other models can do, often with higher quality.", |
75 | | - null, |
76 | | - 2049), |
77 | | - CURIE("curie", |
78 | | - "Very capable, but faster and lower cost than Davinci.", |
79 | | - null, |
80 | | - 2049), |
81 | | - BABBAGE("babbage", |
82 | | - "Capable of straightforward tasks, very fast, and lower cost.", |
83 | | - null, |
84 | | - 2049), |
85 | | - ADA("ada", |
86 | | - "Capable of very simple tasks, usually the fastest model in the GPT-3 series, and lowest cost.", |
87 | | - null, |
88 | | - 2049), |
| 150 | + |
89 | 151 | /* ================================ Claude ================================ */ |
90 | 152 | CLAUDE_2("claude-2", |
91 | 153 | null, |
92 | 154 | null, |
93 | 155 | Integer.MAX_VALUE), |
| 156 | + |
94 | 157 | CLAUDE_INSTANT_1("claude-instant-1", |
95 | 158 | null, |
96 | 159 | null, |
|
0 commit comments