Skip to content

Commit c07f8cd

Browse files
committed
add attributes
1 parent 71caa9c commit c07f8cd

File tree

3 files changed

+25
-10
lines changed

3 files changed

+25
-10
lines changed

veadk/tracing/telemetry/attributes/extractors/common_attributes_extractors.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -62,8 +62,10 @@ def llm_openinference_instrumentation_veadk(**kwargs) -> str:
6262
"gen_ai.user.id": common_gen_ai_user_id, # APMPlus required
6363
"gen_ai.session.id": common_gen_ai_session_id, # APMPlus required
6464
"agent_name": common_gen_ai_agent_name, # CozeLoop required
65+
"agent.name": common_gen_ai_agent_name, # TLS required
6566
"app_name": common_gen_ai_app_name, # CozeLoop required
66-
"user.id": common_gen_ai_user_id, # CozeLoop required
67-
"session.id": common_gen_ai_session_id, # CozeLoop required
67+
"app.name": common_gen_ai_app_name, # TLS required
68+
"user.id": common_gen_ai_user_id, # CozeLoop / TLS required
69+
"session.id": common_gen_ai_session_id, # CozeLoop / TLS required
6870
"cozeloop.report.source": common_cozeloop_report_source, # CozeLoop required
6971
}

veadk/tracing/telemetry/attributes/extractors/llm_attributes_extractors.py

Lines changed: 16 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -360,25 +360,33 @@ def llm_gen_ai_choice(params: LLMAttributesParams) -> ExtractorResponse:
360360

361361

362362
LLM_ATTRIBUTES = {
363+
# ===== request attributes =====
363364
"gen_ai.request.model": llm_gen_ai_request_model,
364365
"gen_ai.request.type": llm_gen_ai_request_type,
365-
"gen_ai.response.model": llm_gen_ai_response_model,
366366
"gen_ai.request.max_tokens": llm_gen_ai_request_max_tokens,
367367
"gen_ai.request.temperature": llm_gen_ai_request_temperature,
368368
"gen_ai.request.top_p": llm_gen_ai_request_top_p,
369-
"gen_ai.prompt": llm_gen_ai_prompt,
370-
"gen_ai.completion": llm_gen_ai_completion,
369+
# ===== response attributes =====
370+
"gen_ai.response.model": llm_gen_ai_response_model,
371371
"gen_ai.response.stop_reason": llm_gen_ai_response_stop_reason,
372372
"gen_ai.response.finish_reason": llm_gen_ai_response_finish_reason,
373-
"gen_ai.usage.input_tokens": llm_gen_ai_usage_input_tokens,
374-
"gen_ai.usage.output_tokens": llm_gen_ai_usage_output_tokens,
375-
"gen_ai.usage.total_tokens": llm_gen_ai_usage_total_tokens,
376-
"gen_ai.usage.cache_creation_input_tokens": llm_gen_ai_usage_cache_creation_input_tokens,
377-
"gen_ai.usage.cache_read_input_tokens": llm_gen_ai_usage_cache_read_input_tokens,
373+
# ===== streaming =====
378374
"gen_ai.is_streaming": llm_gen_ai_is_streaming,
375+
# ===== span type =====
379376
"gen_ai.operation.name": llm_gen_ai_operation_name,
377+
# ===== inputs and outputs =====
378+
# events
380379
"gen_ai.system.message": llm_gen_ai_system_message,
381380
"gen_ai.user.message": llm_gen_ai_user_message,
382381
"gen_ai.assistant.message": llm_gen_ai_assistant_message,
383382
"gen_ai.choice": llm_gen_ai_choice,
383+
# attributes
384+
"gen_ai.prompt": llm_gen_ai_prompt,
385+
"gen_ai.completion": llm_gen_ai_completion,
386+
# ===== usage =====
387+
"gen_ai.usage.input_tokens": llm_gen_ai_usage_input_tokens,
388+
"gen_ai.usage.output_tokens": llm_gen_ai_usage_output_tokens,
389+
"gen_ai.usage.total_tokens": llm_gen_ai_usage_total_tokens,
390+
"gen_ai.usage.cache_creation_input_tokens": llm_gen_ai_usage_cache_creation_input_tokens,
391+
"gen_ai.usage.cache_read_input_tokens": llm_gen_ai_usage_cache_read_input_tokens,
384392
}

veadk/tracing/telemetry/attributes/extractors/tool_attributes_extractors.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -48,6 +48,10 @@ def tool_cozeloop_input(params: ToolAttributesParams) -> ExtractorResponse:
4848
return ExtractorResponse(content=json.dumps(tool_input) or "<unknown_tool_input>")
4949

5050

51+
def tool_gen_ai_tool_name(params: ToolAttributesParams) -> ExtractorResponse:
52+
return ExtractorResponse(content=params.tool.name or "<unknown_tool_name>")
53+
54+
5155
def tool_cozeloop_output(params: ToolAttributesParams) -> ExtractorResponse:
5256
function_response = params.function_response_event.get_function_responses()[0]
5357
tool_output = {
@@ -60,6 +64,7 @@ def tool_cozeloop_output(params: ToolAttributesParams) -> ExtractorResponse:
6064

6165
TOOL_ATTRIBUTES = {
6266
"gen_ai.operation.name": tool_gen_ai_operation_name,
67+
"gen_ai.tool.name": tool_gen_ai_tool_name, # TLS required
6368
"cozeloop.input": tool_cozeloop_input, # CozeLoop required
6469
"cozeloop.output": tool_cozeloop_output, # CozeLoop required
6570
}

0 commit comments

Comments
 (0)