|
18 | 18 |
|
19 | 19 | import openai
|
20 | 20 | import pytest
|
21 |
| -from opentelemetry.instrumentation.openai.helpers import GEN_AI_REQUEST_ENCODING_FORMAT |
| 21 | +from opentelemetry.instrumentation.openai.helpers import GEN_AI_REQUEST_ENCODING_FORMATS |
22 | 22 | from opentelemetry.trace import SpanKind, StatusCode
|
23 | 23 | from opentelemetry.semconv._incubating.attributes.gen_ai_attributes import (
|
24 | 24 | GEN_AI_OPERATION_NAME,
|
@@ -117,7 +117,7 @@ def test_all_the_client_options(provider_str, model, input_tokens, duration, tra
|
117 | 117 | GEN_AI_REQUEST_MODEL: model,
|
118 | 118 | GEN_AI_SYSTEM: "openai",
|
119 | 119 | GEN_AI_RESPONSE_MODEL: model,
|
120 |
| - GEN_AI_REQUEST_ENCODING_FORMAT: "float", |
| 120 | + GEN_AI_REQUEST_ENCODING_FORMATS: ("float",), |
121 | 121 | GEN_AI_USAGE_INPUT_TOKENS: input_tokens,
|
122 | 122 | SERVER_ADDRESS: provider.server_address,
|
123 | 123 | SERVER_PORT: provider.server_port,
|
@@ -159,7 +159,7 @@ def test_all_the_client_options_integration(provider_str, model, trace_exporter,
|
159 | 159 | GEN_AI_REQUEST_MODEL: model,
|
160 | 160 | GEN_AI_SYSTEM: "openai",
|
161 | 161 | GEN_AI_RESPONSE_MODEL: model,
|
162 |
| - GEN_AI_REQUEST_ENCODING_FORMAT: "float", |
| 162 | + GEN_AI_REQUEST_ENCODING_FORMATS: ("float",), |
163 | 163 | GEN_AI_USAGE_INPUT_TOKENS: response.usage.prompt_tokens,
|
164 | 164 | SERVER_ADDRESS: provider.server_address,
|
165 | 165 | SERVER_PORT: provider.server_port,
|
@@ -312,7 +312,7 @@ async def test_async_all_the_client_options(
|
312 | 312 | GEN_AI_REQUEST_MODEL: model,
|
313 | 313 | GEN_AI_SYSTEM: "openai",
|
314 | 314 | GEN_AI_RESPONSE_MODEL: model,
|
315 |
| - GEN_AI_REQUEST_ENCODING_FORMAT: "float", |
| 315 | + GEN_AI_REQUEST_ENCODING_FORMATS: ("float",), |
316 | 316 | GEN_AI_USAGE_INPUT_TOKENS: input_tokens,
|
317 | 317 | SERVER_ADDRESS: provider.server_address,
|
318 | 318 | SERVER_PORT: provider.server_port,
|
@@ -355,7 +355,7 @@ async def test_async_all_the_client_options_integration(provider_str, model, tra
|
355 | 355 | GEN_AI_REQUEST_MODEL: model,
|
356 | 356 | GEN_AI_SYSTEM: "openai",
|
357 | 357 | GEN_AI_RESPONSE_MODEL: model,
|
358 |
| - GEN_AI_REQUEST_ENCODING_FORMAT: "float", |
| 358 | + GEN_AI_REQUEST_ENCODING_FORMATS: ("float",), |
359 | 359 | GEN_AI_USAGE_INPUT_TOKENS: response.usage.prompt_tokens,
|
360 | 360 | SERVER_ADDRESS: provider.server_address,
|
361 | 361 | SERVER_PORT: provider.server_port,
|
|
0 commit comments