|
14 | 14 |
|
15 | 15 | from __future__ import annotations |
16 | 16 |
|
| 17 | +import re |
17 | 18 | from dataclasses import dataclass |
18 | 19 | from os import environ |
19 | 20 | from typing import ( |
20 | 21 | TYPE_CHECKING, |
21 | | - Dict, |
22 | | - List, |
23 | 22 | Mapping, |
24 | 23 | Optional, |
25 | | - TypedDict, |
26 | | - cast, |
| 24 | + Sequence, |
27 | 25 | ) |
28 | 26 |
|
29 | 27 | from opentelemetry.semconv._incubating.attributes import ( |
|
32 | 30 | from opentelemetry.util.types import AttributeValue |
33 | 31 |
|
34 | 32 | if TYPE_CHECKING: |
35 | | - from vertexai.generative_models import Tool, ToolConfig |
36 | | - from vertexai.generative_models._generative_models import ( |
37 | | - ContentsType, |
38 | | - GenerationConfigType, |
39 | | - SafetySettingsType, |
40 | | - _GenerativeModel, |
41 | | - ) |
| 33 | + from google.cloud.aiplatform_v1.types import content, tool |
42 | 34 |
|
43 | 35 |
|
44 | 36 | @dataclass(frozen=True) |
45 | 37 | class GenerateContentParams: |
46 | | - contents: ContentsType |
47 | | - generation_config: Optional[GenerationConfigType] |
48 | | - safety_settings: Optional[SafetySettingsType] |
49 | | - tools: Optional[List["Tool"]] |
50 | | - tool_config: Optional["ToolConfig"] |
51 | | - labels: Optional[Dict[str, str]] |
52 | | - stream: bool |
53 | | - |
54 | | - |
55 | | -class GenerationConfigDict(TypedDict, total=False): |
56 | | - temperature: Optional[float] |
57 | | - top_p: Optional[float] |
58 | | - top_k: Optional[int] |
59 | | - max_output_tokens: Optional[int] |
60 | | - stop_sequences: Optional[List[str]] |
61 | | - presence_penalty: Optional[float] |
62 | | - frequency_penalty: Optional[float] |
63 | | - seed: Optional[int] |
64 | | - # And more fields which aren't needed yet |
| 38 | + model: str |
| 39 | + contents: Optional[Sequence[content.Content]] = None |
| 40 | + system_instruction: Optional[content.Content | None] = None |
| 41 | + tools: Optional[Sequence[tool.Tool]] = None |
| 42 | + tool_config: Optional[tool.ToolConfig] = None |
| 43 | + labels: Optional[Mapping[str, str]] = None |
| 44 | + safety_settings: Optional[Sequence[content.SafetySetting]] = None |
| 45 | + generation_config: Optional[content.GenerationConfig] = None |
65 | 46 |
|
66 | 47 |
|
67 | 48 | def get_genai_request_attributes( |
68 | | - instance: _GenerativeModel, |
69 | 49 | params: GenerateContentParams, |
70 | 50 | operation_name: GenAIAttributes.GenAiOperationNameValues = GenAIAttributes.GenAiOperationNameValues.CHAT, |
71 | 51 | ): |
72 | | - model = _get_model_name(instance) |
73 | | - generation_config = _get_generation_config(instance, params) |
74 | | - attributes = { |
| 52 | + model = _get_model_name(params.model) |
| 53 | + generation_config = params.generation_config |
| 54 | + attributes: dict[str, AttributeValue] = { |
75 | 55 | GenAIAttributes.GEN_AI_OPERATION_NAME: operation_name.value, |
76 | 56 | GenAIAttributes.GEN_AI_SYSTEM: GenAIAttributes.GenAiSystemValues.VERTEX_AI.value, |
77 | 57 | GenAIAttributes.GEN_AI_REQUEST_MODEL: model, |
78 | | - GenAIAttributes.GEN_AI_REQUEST_TEMPERATURE: generation_config.get( |
79 | | - "temperature" |
80 | | - ), |
81 | | - GenAIAttributes.GEN_AI_REQUEST_TOP_P: generation_config.get("top_p"), |
82 | | - GenAIAttributes.GEN_AI_REQUEST_MAX_TOKENS: generation_config.get( |
83 | | - "max_output_tokens" |
84 | | - ), |
85 | | - GenAIAttributes.GEN_AI_REQUEST_PRESENCE_PENALTY: generation_config.get( |
86 | | - "presence_penalty" |
87 | | - ), |
88 | | - GenAIAttributes.GEN_AI_REQUEST_FREQUENCY_PENALTY: generation_config.get( |
89 | | - "frequency_penalty" |
90 | | - ), |
91 | | - GenAIAttributes.GEN_AI_OPENAI_REQUEST_SEED: generation_config.get( |
92 | | - "seed" |
93 | | - ), |
94 | | - GenAIAttributes.GEN_AI_REQUEST_STOP_SEQUENCES: generation_config.get( |
95 | | - "stop_sequences" |
96 | | - ), |
97 | 58 | } |
98 | 59 |
|
99 | | - # filter out None values |
100 | | - return {k: v for k, v in attributes.items() if v is not None} |
101 | | - |
102 | | - |
103 | | -def _get_generation_config( |
104 | | - instance: _GenerativeModel, |
105 | | - params: GenerateContentParams, |
106 | | -) -> GenerationConfigDict: |
107 | | - generation_config = params.generation_config or instance._generation_config |
108 | | - if generation_config is None: |
109 | | - return {} |
110 | | - if isinstance(generation_config, dict): |
111 | | - return cast(GenerationConfigDict, generation_config) |
112 | | - return cast(GenerationConfigDict, generation_config.to_dict()) |
113 | | - |
114 | | - |
115 | | -_RESOURCE_PREFIX = "publishers/google/models/" |
116 | | - |
| 60 | + if not generation_config: |
| 61 | + return attributes |
| 62 | + |
| 63 | + # Check for optional fields |
| 64 | + # https://proto-plus-python.readthedocs.io/en/stable/fields.html#optional-fields |
| 65 | + if "temperature" in generation_config: |
| 66 | + attributes[GenAIAttributes.GEN_AI_REQUEST_TEMPERATURE] = ( |
| 67 | + generation_config.temperature |
| 68 | + ) |
| 69 | + if "top_p" in generation_config: |
| 70 | + attributes[GenAIAttributes.GEN_AI_REQUEST_TOP_P] = ( |
| 71 | + generation_config.top_p |
| 72 | + ) |
| 73 | + if "max_output_tokens" in generation_config: |
| 74 | + attributes[GenAIAttributes.GEN_AI_REQUEST_MAX_TOKENS] = ( |
| 75 | + generation_config.max_output_tokens |
| 76 | + ) |
| 77 | + if "presence_penalty" in generation_config: |
| 78 | + attributes[GenAIAttributes.GEN_AI_REQUEST_PRESENCE_PENALTY] = ( |
| 79 | + generation_config.presence_penalty |
| 80 | + ) |
| 81 | + if "frequency_penalty" in generation_config: |
| 82 | + attributes[GenAIAttributes.GEN_AI_REQUEST_FREQUENCY_PENALTY] = ( |
| 83 | + generation_config.frequency_penalty |
| 84 | + ) |
| 85 | + if "seed" in generation_config: |
| 86 | + attributes[GenAIAttributes.GEN_AI_OPENAI_REQUEST_SEED] = ( |
| 87 | + generation_config.seed |
| 88 | + ) |
| 89 | + if "stop_sequences" in generation_config: |
| 90 | + attributes[GenAIAttributes.GEN_AI_REQUEST_STOP_SEQUENCES] = ( |
| 91 | + generation_config.stop_sequences |
| 92 | + ) |
| 93 | + |
| 94 | + return attributes |
| 95 | + |
| 96 | + |
| 97 | +_MODEL_STRIP_RE = re.compile( |
| 98 | + r"^projects/(.*)/locations/(.*)/publishers/google/models/" |
| 99 | +) |
117 | 100 |
|
118 | | -def _get_model_name(instance: _GenerativeModel) -> str: |
119 | | - model_name = instance._model_name |
120 | 101 |
|
121 | | - # Can use str.removeprefix() once 3.8 is dropped |
122 | | - if model_name.startswith(_RESOURCE_PREFIX): |
123 | | - model_name = model_name[len(_RESOURCE_PREFIX) :] |
124 | | - return model_name |
| 102 | +def _get_model_name(model: str) -> str: |
| 103 | + return _MODEL_STRIP_RE.sub("", model) |
125 | 104 |
|
126 | 105 |
|
127 | 106 | OTEL_INSTRUMENTATION_GENAI_CAPTURE_MESSAGE_CONTENT = ( |
|
0 commit comments