|
19 | 19 | "amazon.titan-text-express-v1": '{ "inputText": "%s", "textGenerationConfig": {"temperature": %f, "maxTokenCount": %d }}', |
20 | 20 | "ai21.j2-mid-v1": '{"prompt": "%s", "temperature": %f, "maxTokens": %d}', |
21 | 21 | "anthropic.claude-instant-v1": '{"prompt": "Human: %s Assistant:", "temperature": %f, "max_tokens_to_sample": %d}', |
| 22 | + "anthropic.claude-3-sonnet-20240229-v1:0": '{"anthropic_version": "bedrock-2023-05-31", "messages": [{"role": "user", "content": "%s"}], "temperature": %f, "max_tokens": %d}', |
22 | 23 | "cohere.command-text-v14": '{"prompt": "%s", "temperature": %f, "max_tokens": %d}', |
23 | 24 | "meta.llama2-13b-chat-v1": '{"prompt": "%s", "temperature": %f, "max_gen_len": %d}', |
24 | 25 | "mistral.mistral-7b-instruct-v0:2": '{"prompt": "<s>[INST] %s [/INST]", "temperature": %f, "max_tokens": %d}', |
|
262 | 263 | }, |
263 | 264 | ), |
264 | 265 | ], |
| 266 | + "anthropic.claude-3-sonnet-20240229-v1:0": [ |
| 267 | + ( |
| 268 | + {"type": "LlmChatCompletionSummary"}, |
| 269 | + { |
| 270 | + "id": None, # UUID that varies with each run |
| 271 | + "llm.conversation_id": "my-awesome-id", |
| 272 | + "llm.foo": "bar", |
| 273 | + "span_id": None, |
| 274 | + "trace_id": "trace-id", |
| 275 | + "request_id": "ab38295d-df9c-4141-8173-38221651bf46", |
| 276 | + "duration": None, # Response time varies each test run |
| 277 | + "request.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 278 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 279 | + "request.temperature": 0.7, |
| 280 | + "request.max_tokens": 100, |
| 281 | + "response.choices.finish_reason": "end_turn", |
| 282 | + "vendor": "bedrock", |
| 283 | + "ingest_source": "Python", |
| 284 | + "response.number_of_messages": 2, |
| 285 | + }, |
| 286 | + ), |
| 287 | + ( |
| 288 | + {"type": "LlmChatCompletionMessage"}, |
| 289 | + { |
| 290 | + "id": None, # UUID that varies with each run |
| 291 | + "llm.conversation_id": "my-awesome-id", |
| 292 | + "llm.foo": "bar", |
| 293 | + "request_id": "ab38295d-df9c-4141-8173-38221651bf46", |
| 294 | + "span_id": None, |
| 295 | + "trace_id": "trace-id", |
| 296 | + "content": "What is 212 degrees Fahrenheit converted to Celsius?", |
| 297 | + "role": "user", |
| 298 | + "completion_id": None, |
| 299 | + "sequence": 0, |
| 300 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 301 | + "vendor": "bedrock", |
| 302 | + "ingest_source": "Python", |
| 303 | + }, |
| 304 | + ), |
| 305 | + ( |
| 306 | + {"type": "LlmChatCompletionMessage"}, |
| 307 | + { |
| 308 | + "id": None, # UUID that varies with each run |
| 309 | + "llm.conversation_id": "my-awesome-id", |
| 310 | + "llm.foo": "bar", |
| 311 | + "request_id": "ab38295d-df9c-4141-8173-38221651bf46", |
| 312 | + "span_id": None, |
| 313 | + "trace_id": "trace-id", |
| 314 | + "content": "212 degrees Fahrenheit equals 100 degrees Celsius. This is the boiling point of water at standard atmospheric pressure.", |
| 315 | + "role": "assistant", |
| 316 | + "completion_id": None, |
| 317 | + "sequence": 1, |
| 318 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 319 | + "vendor": "bedrock", |
| 320 | + "ingest_source": "Python", |
| 321 | + "is_response": True, |
| 322 | + }, |
| 323 | + ), |
| 324 | + ], |
265 | 325 | "cohere.command-text-v14": [ |
266 | 326 | ( |
267 | 327 | {"type": "LlmChatCompletionSummary"}, |
|
555 | 615 | }, |
556 | 616 | ), |
557 | 617 | ], |
| 618 | + "anthropic.claude-3-sonnet-20240229-v1:0": [ |
| 619 | + ( |
| 620 | + {"type": "LlmChatCompletionSummary"}, |
| 621 | + { |
| 622 | + "id": None, # UUID that varies with each run |
| 623 | + "llm.conversation_id": "my-awesome-id", |
| 624 | + "llm.foo": "bar", |
| 625 | + "span_id": None, |
| 626 | + "trace_id": "trace-id", |
| 627 | + "request_id": "e8fc1dd7-3d1e-42c6-9c58-535cae563bff", |
| 628 | + "duration": None, # Response time varies each test run |
| 629 | + "request.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 630 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 631 | + "vendor": "bedrock", |
| 632 | + "ingest_source": "Python", |
| 633 | + "response.number_of_messages": 2, |
| 634 | + }, |
| 635 | + ), |
| 636 | + ( |
| 637 | + {"type": "LlmChatCompletionMessage"}, |
| 638 | + { |
| 639 | + "id": None, # UUID that varies with each run |
| 640 | + "llm.conversation_id": "my-awesome-id", |
| 641 | + "llm.foo": "bar", |
| 642 | + "request_id": "e8fc1dd7-3d1e-42c6-9c58-535cae563bff", |
| 643 | + "span_id": None, |
| 644 | + "trace_id": "trace-id", |
| 645 | + "content": "The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.\n\nCurrent conversation:\n\nHuman: Hi there!\nAI:", |
| 646 | + "role": "user", |
| 647 | + "completion_id": None, |
| 648 | + "sequence": 0, |
| 649 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 650 | + "vendor": "bedrock", |
| 651 | + "ingest_source": "Python", |
| 652 | + }, |
| 653 | + ), |
| 654 | + ( |
| 655 | + {"type": "LlmChatCompletionMessage"}, |
| 656 | + { |
| 657 | + "id": None, # UUID that varies with each run |
| 658 | + "llm.conversation_id": "my-awesome-id", |
| 659 | + "llm.foo": "bar", |
| 660 | + "request_id": "e8fc1dd7-3d1e-42c6-9c58-535cae563bff", |
| 661 | + "span_id": None, |
| 662 | + "trace_id": "trace-id", |
| 663 | + "content": "I'm ready for a friendly conversation! I'll share specific details when I can, and if I don't know something, I'll be straightforward about that. What would you like to talk about today?", |
| 664 | + "role": "assistant", |
| 665 | + "completion_id": None, |
| 666 | + "sequence": 1, |
| 667 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 668 | + "vendor": "bedrock", |
| 669 | + "ingest_source": "Python", |
| 670 | + "is_response": True, |
| 671 | + }, |
| 672 | + ), |
| 673 | + ], |
558 | 674 | "meta.llama2-13b-chat-v1": [ |
559 | 675 | ( |
560 | 676 | {"type": "LlmChatCompletionSummary"}, |
|
787 | 903 | }, |
788 | 904 | ), |
789 | 905 | ], |
| 906 | + "anthropic.claude-3-sonnet-20240229-v1:0": [ |
| 907 | + ( |
| 908 | + {"type": "LlmChatCompletionSummary"}, |
| 909 | + { |
| 910 | + "id": None, # UUID that varies with each run |
| 911 | + "llm.conversation_id": "my-awesome-id", |
| 912 | + "llm.foo": "bar", |
| 913 | + "span_id": None, |
| 914 | + "trace_id": "trace-id", |
| 915 | + "request_id": "96c7306d-2d60-4629-83e9-dbd6befb0e4e", |
| 916 | + "duration": None, # Response time varies each test run |
| 917 | + "request.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 918 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 919 | + "response.choices.finish_reason": "end_turn", |
| 920 | + "vendor": "bedrock", |
| 921 | + "ingest_source": "Python", |
| 922 | + "response.number_of_messages": 2, |
| 923 | + }, |
| 924 | + ), |
| 925 | + ( |
| 926 | + {"type": "LlmChatCompletionMessage"}, |
| 927 | + { |
| 928 | + "id": None, # UUID that varies with each run |
| 929 | + "llm.conversation_id": "my-awesome-id", |
| 930 | + "llm.foo": "bar", |
| 931 | + "request_id": "96c7306d-2d60-4629-83e9-dbd6befb0e4e", |
| 932 | + "span_id": None, |
| 933 | + "trace_id": "trace-id", |
| 934 | + "content": "The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.\n\nCurrent conversation:\n\nHuman: Hi there!\nAI:", |
| 935 | + "role": "user", |
| 936 | + "completion_id": None, |
| 937 | + "sequence": 0, |
| 938 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 939 | + "vendor": "bedrock", |
| 940 | + "ingest_source": "Python", |
| 941 | + }, |
| 942 | + ), |
| 943 | + ( |
| 944 | + {"type": "LlmChatCompletionMessage"}, |
| 945 | + { |
| 946 | + "id": None, # UUID that varies with each run |
| 947 | + "llm.conversation_id": "my-awesome-id", |
| 948 | + "llm.foo": "bar", |
| 949 | + "request_id": "96c7306d-2d60-4629-83e9-dbd6befb0e4e", |
| 950 | + "span_id": None, |
| 951 | + "trace_id": "trace-id", |
| 952 | + "content": "I'm ready for a friendly conversation! I'll share specific details when I can, and if I don't know something, I'll be straightforward about that. What would you like to talk about today?", |
| 953 | + "role": "assistant", |
| 954 | + "completion_id": None, |
| 955 | + "sequence": 1, |
| 956 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 957 | + "vendor": "bedrock", |
| 958 | + "ingest_source": "Python", |
| 959 | + "is_response": True, |
| 960 | + }, |
| 961 | + ), |
| 962 | + ], |
790 | 963 | "meta.llama2-13b-chat-v1": [ |
791 | 964 | ( |
792 | 965 | {"type": "LlmChatCompletionSummary"}, |
|
1024 | 1197 | }, |
1025 | 1198 | ), |
1026 | 1199 | ], |
| 1200 | + "anthropic.claude-3-sonnet-20240229-v1:0": [ |
| 1201 | + ( |
| 1202 | + {"type": "LlmChatCompletionSummary"}, |
| 1203 | + { |
| 1204 | + "id": None, # UUID that varies with each run |
| 1205 | + "llm.conversation_id": "my-awesome-id", |
| 1206 | + "llm.foo": "bar", |
| 1207 | + "span_id": None, |
| 1208 | + "trace_id": "trace-id", |
| 1209 | + "request_id": "1efe6197-80f9-43a6-89a5-bb536c1b822f", |
| 1210 | + "duration": None, # Response time varies each test run |
| 1211 | + "request.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1212 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1213 | + "request.temperature": 0.7, |
| 1214 | + "request.max_tokens": 100, |
| 1215 | + "vendor": "bedrock", |
| 1216 | + "ingest_source": "Python", |
| 1217 | + "response.number_of_messages": 2, |
| 1218 | + }, |
| 1219 | + ), |
| 1220 | + ( |
| 1221 | + {"type": "LlmChatCompletionMessage"}, |
| 1222 | + { |
| 1223 | + "id": None, # UUID that varies with each run |
| 1224 | + "llm.conversation_id": "my-awesome-id", |
| 1225 | + "llm.foo": "bar", |
| 1226 | + "request_id": "1efe6197-80f9-43a6-89a5-bb536c1b822f", |
| 1227 | + "span_id": None, |
| 1228 | + "trace_id": "trace-id", |
| 1229 | + "content": "What is 212 degrees Fahrenheit converted to Celsius?", |
| 1230 | + "role": "user", |
| 1231 | + "completion_id": None, |
| 1232 | + "sequence": 0, |
| 1233 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1234 | + "vendor": "bedrock", |
| 1235 | + "ingest_source": "Python", |
| 1236 | + }, |
| 1237 | + ), |
| 1238 | + ( |
| 1239 | + {"type": "LlmChatCompletionMessage"}, |
| 1240 | + { |
| 1241 | + "id": None, # UUID that varies with each run |
| 1242 | + "llm.conversation_id": "my-awesome-id", |
| 1243 | + "llm.foo": "bar", |
| 1244 | + "request_id": "1efe6197-80f9-43a6-89a5-bb536c1b822f", |
| 1245 | + "span_id": None, |
| 1246 | + "trace_id": "trace-id", |
| 1247 | + "content": "212 degrees Fahrenheit is equivalent to 100 degrees Celsius.\n\nThis is the boiling point of water at standard atmospheric pressure.", |
| 1248 | + "role": "assistant", |
| 1249 | + "completion_id": None, |
| 1250 | + "sequence": 1, |
| 1251 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1252 | + "vendor": "bedrock", |
| 1253 | + "ingest_source": "Python", |
| 1254 | + "is_response": True, |
| 1255 | + }, |
| 1256 | + ), |
| 1257 | + ], |
1027 | 1258 | "cohere.command-text-v14": [ |
1028 | 1259 | ( |
1029 | 1260 | {"type": "LlmChatCompletionSummary"}, |
|
1326 | 1557 | }, |
1327 | 1558 | ), |
1328 | 1559 | ], |
| 1560 | + "anthropic.claude-3-sonnet-20240229-v1:0": [ |
| 1561 | + ( |
| 1562 | + {"type": "LlmChatCompletionSummary"}, |
| 1563 | + { |
| 1564 | + "id": None, # UUID that varies with each run |
| 1565 | + "llm.conversation_id": "my-awesome-id", |
| 1566 | + "llm.foo": "bar", |
| 1567 | + "span_id": None, |
| 1568 | + "trace_id": "trace-id", |
| 1569 | + "request_id": "282ba076-576f-46aa-a2e6-680392132e87", |
| 1570 | + "duration": None, # Response time varies each test run |
| 1571 | + "request.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1572 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1573 | + "request.temperature": 0.7, |
| 1574 | + "request.max_tokens": 100, |
| 1575 | + "vendor": "bedrock", |
| 1576 | + "ingest_source": "Python", |
| 1577 | + "response.number_of_messages": 1, |
| 1578 | + "error": True, |
| 1579 | + }, |
| 1580 | + ), |
| 1581 | + ( |
| 1582 | + {"type": "LlmChatCompletionMessage"}, |
| 1583 | + { |
| 1584 | + "id": None, # UUID that varies with each run |
| 1585 | + "llm.conversation_id": "my-awesome-id", |
| 1586 | + "llm.foo": "bar", |
| 1587 | + "request_id": "282ba076-576f-46aa-a2e6-680392132e87", |
| 1588 | + "span_id": None, |
| 1589 | + "trace_id": "trace-id", |
| 1590 | + "content": "Invalid Token", |
| 1591 | + "role": "user", |
| 1592 | + "completion_id": None, |
| 1593 | + "sequence": 0, |
| 1594 | + "response.model": "anthropic.claude-3-sonnet-20240229-v1:0", |
| 1595 | + "vendor": "bedrock", |
| 1596 | + "ingest_source": "Python", |
| 1597 | + }, |
| 1598 | + ), |
| 1599 | + ], |
1329 | 1600 | "cohere.command-text-v14": [ |
1330 | 1601 | ( |
1331 | 1602 | {"type": "LlmChatCompletionSummary"}, |
|
0 commit comments