Skip to content
16 changes: 13 additions & 3 deletions newrelic/hooks/mlmodel_openai.py
Original file line number Diff line number Diff line change
Expand Up @@ -442,6 +442,7 @@ def _handle_completion_success(transaction, linking_metadata, completion_id, kwa
# The function trace will be exited when in the final iteration of the response
# generator.
return_val._nr_ft = ft
return_val._nr_metadata = linking_metadata
return_val._nr_openai_attrs = getattr(return_val, "_nr_openai_attrs", {})
return_val._nr_openai_attrs["messages"] = kwargs.get("messages", [])
return_val._nr_openai_attrs["temperature"] = kwargs.get("temperature")
Expand Down Expand Up @@ -492,10 +493,14 @@ def _record_completion_success(transaction, linking_metadata, completion_id, kwa
response_model = kwargs.get("response.model")
response_id = kwargs.get("id")
output_message_list = []
finish_reason = None
finish_reason = kwargs.get("finish_reason")
if "content" in kwargs:
output_message_list = [{"content": kwargs.get("content"), "role": kwargs.get("role")}]
finish_reason = kwargs.get("finish_reason")
# When tools are involved, the content key may hold an empty string which we do not want to report
# In this case, the content we are interested in capturing will already be covered in the input_message_list
# We empty out the output_message_list so that we do not report an empty message
if "tool_call" in finish_reason and not kwargs.get("content"):
output_message_list = []
request_model = kwargs.get("model") or kwargs.get("engine")

request_id = response_headers.get("x-request-id")
Expand Down Expand Up @@ -765,7 +770,10 @@ def _record_stream_chunk(self, return_val):

def _record_events_on_stop_iteration(self, transaction):
if hasattr(self, "_nr_ft"):
linking_metadata = get_trace_linking_metadata()
# We first check for our saved linking metadata before making a new call to get_trace_linking_metadata
# Directly calling get_trace_linking_metadata() causes the incorrect span ID to be captured and associated with the LLM call
# This leads to incorrect linking of the LLM call in the UI
linking_metadata = self._nr_metadata or get_trace_linking_metadata()
self._nr_ft.__exit__(None, None, None)
try:
openai_attrs = getattr(self, "_nr_openai_attrs", {})
Expand Down Expand Up @@ -872,6 +880,8 @@ def set_attrs_on_generator_proxy(proxy, instance):
proxy._nr_response_headers = instance._nr_response_headers
if hasattr(instance, "_nr_openai_attrs"):
proxy._nr_openai_attrs = instance._nr_openai_attrs
if hasattr(instance, "_nr_metadata"):
proxy._nr_metadata = instance._nr_metadata


def wrap_engine_api_resource_create_sync(wrapped, instance, args, kwargs):
Expand Down
Loading