@@ -1013,9 +1013,6 @@ def test_langchain_llm_exception_captured(sentry_init, capture_events):
10131013
10141014def test_langchain_different_exception_types (sentry_init , capture_events ):
10151015 """Test that different exception types are properly captured."""
1016- global llm_type
1017- llm_type = "openai-chat"
1018-
10191016 exception_types = [
10201017 (ValueError , "Invalid parameter" ),
10211018 (TypeError , "Type mismatch" ),
@@ -1066,9 +1063,6 @@ def test_langchain_different_exception_types(sentry_init, capture_events):
10661063
10671064def test_langchain_exception_with_span_context (sentry_init , capture_events ):
10681065 """Test that exception events include proper span context."""
1069- global llm_type
1070- llm_type = "openai-chat"
1071-
10721066 sentry_init (
10731067 integrations = [LangchainIntegration (include_prompts = True )],
10741068 traces_sample_rate = 1.0 ,
@@ -1123,9 +1117,6 @@ def test_langchain_exception_with_span_context(sentry_init, capture_events):
11231117
11241118def test_langchain_tool_execution_error (sentry_init , capture_events ):
11251119 """Test that exceptions during tool execution are properly captured."""
1126- global llm_type
1127- llm_type = "openai-chat"
1128-
11291120 sentry_init (
11301121 integrations = [LangchainIntegration (include_prompts = True )],
11311122 traces_sample_rate = 1.0 ,
@@ -1198,9 +1189,6 @@ def failing_tool(word: str) -> int:
11981189
11991190def test_langchain_exception_span_cleanup (sentry_init , capture_events ):
12001191 """Test that spans are properly cleaned up even when exceptions occur."""
1201- global llm_type
1202- llm_type = "openai-chat"
1203-
12041192 sentry_init (
12051193 integrations = [LangchainIntegration (include_prompts = True )],
12061194 traces_sample_rate = 1.0 ,
0 commit comments