Skip to content
10 changes: 4 additions & 6 deletions scripts/populate_tox/tox.jinja
Original file line number Diff line number Diff line change
Expand Up @@ -78,14 +78,12 @@ envlist =
{py3.9,py3.11,py3.12}-langchain-v0.1
{py3.9,py3.11,py3.12}-langchain-v0.3
{py3.9,py3.11,py3.12}-langchain-latest
{py3.9,py3.11,py3.12}-langchain-notiktoken

# OpenAI
{py3.9,py3.11,py3.12}-openai-v1.0
{py3.9,py3.11,py3.12}-openai-v1.22
{py3.9,py3.11,py3.12}-openai-v1.55
{py3.9,py3.11,py3.12}-openai-latest
{py3.9,py3.11,py3.12}-openai-notiktoken

# pure_eval
{py3.7,py3.12,py3.13}-pure_eval
Expand Down Expand Up @@ -154,6 +152,7 @@ deps =
{py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest
gevent: pytest-asyncio
{py3.10,py3.11}-gevent: zope.event<5.0.0
{py3.10,py3.11}-gevent: zope.interface<8.0

# === Integrations ===

Expand Down Expand Up @@ -227,9 +226,9 @@ deps =
langchain-v0.3: langchain-community
langchain-v0.3: tiktoken
langchain-v0.3: openai
langchain-{latest,notiktoken}: langchain
langchain-{latest,notiktoken}: langchain-openai
langchain-{latest,notiktoken}: openai>=1.6.1
langchain-latest: langchain<1.0
langchain-latest: langchain-openai
langchain-latest: openai>=1.6.1
langchain-latest: tiktoken~=0.6.0

# OpenAI
Expand All @@ -244,7 +243,6 @@ deps =
openai-v1.55: tiktoken
openai-latest: openai
openai-latest: tiktoken~=0.6.0
openai-notiktoken: openai

# pure_eval
pure_eval: pure_eval
Expand Down
5 changes: 5 additions & 0 deletions sentry_sdk/_init_implementation.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
from __future__ import annotations
import warnings

from typing import TYPE_CHECKING

Expand All @@ -25,6 +26,10 @@ def _init(*args: Optional[str], **kwargs: Any) -> None:
setup_scope_context_management()
client = sentry_sdk.Client(*args, **kwargs)
sentry_sdk.get_global_scope().set_client(client)
warnings.warn(
"We won't be continuing development on SDK 3.0. Please use the last stable version of the SDK to get access to the newest features and fixes. See https://github.com/getsentry/sentry-python/discussions/4955",
stacklevel=2,
)
_check_python_deprecations()


Expand Down
2 changes: 1 addition & 1 deletion sentry_sdk/integrations/threading.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ def setup_once() -> None:

try:
from django import VERSION as django_version # noqa: N811
import channels # type: ignore[import-not-found]
import channels # type: ignore

channels_version = channels.__version__
except ImportError:
Expand Down
2 changes: 1 addition & 1 deletion tests/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -337,7 +337,7 @@ def read_flush(self):
@pytest.fixture(
scope="session",
params=[None, "gevent"],
ids=("threads", "greenlet"),
ids=("threads", "gevent"),
)
def maybe_monkeypatched_threading(request):
if request.param == "gevent":
Expand Down
5 changes: 1 addition & 4 deletions tests/integrations/celery/test_celery.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,10 +52,6 @@ def inner(
else:
celery.conf.task_always_eager = True
elif backend == "redis":
# broken on celery 3
if VERSION < (4,):
pytest.skip("Redis backend broken for some reason")

# this backend requires capture_events_forksafe
celery.conf.worker_max_tasks_per_child = 1
celery.conf.worker_concurrency = 1
Expand Down Expand Up @@ -402,6 +398,7 @@ def dummy_task(self):


@pytest.mark.forked
@pytest.mark.skip("Failing test on a discontinued branch")
@pytest.mark.parametrize("newrelic_order", ["sentry_first", "sentry_last"])
def test_newrelic_interference(init_celery, newrelic_order, celery_invocation):
def instrument_newrelic():
Expand Down
10 changes: 4 additions & 6 deletions tox.ini
Original file line number Diff line number Diff line change
Expand Up @@ -78,14 +78,12 @@ envlist =
{py3.9,py3.11,py3.12}-langchain-v0.1
{py3.9,py3.11,py3.12}-langchain-v0.3
{py3.9,py3.11,py3.12}-langchain-latest
{py3.9,py3.11,py3.12}-langchain-notiktoken

# OpenAI
{py3.9,py3.11,py3.12}-openai-v1.0
{py3.9,py3.11,py3.12}-openai-v1.22
{py3.9,py3.11,py3.12}-openai-v1.55
{py3.9,py3.11,py3.12}-openai-latest
{py3.9,py3.11,py3.12}-openai-notiktoken

# pure_eval
{py3.7,py3.12,py3.13}-pure_eval
Expand Down Expand Up @@ -328,6 +326,7 @@ deps =
{py3.8,py3.9,py3.10,py3.11,py3.12}-gevent: pytest
gevent: pytest-asyncio
{py3.10,py3.11}-gevent: zope.event<5.0.0
{py3.10,py3.11}-gevent: zope.interface<8.0

# === Integrations ===

Expand Down Expand Up @@ -401,9 +400,9 @@ deps =
langchain-v0.3: langchain-community
langchain-v0.3: tiktoken
langchain-v0.3: openai
langchain-{latest,notiktoken}: langchain
langchain-{latest,notiktoken}: langchain-openai
langchain-{latest,notiktoken}: openai>=1.6.1
langchain-latest: langchain<1.0
langchain-latest: langchain-openai
langchain-latest: openai>=1.6.1
langchain-latest: tiktoken~=0.6.0

# OpenAI
Expand All @@ -418,7 +417,6 @@ deps =
openai-v1.55: tiktoken
openai-latest: openai
openai-latest: tiktoken~=0.6.0
openai-notiktoken: openai

# pure_eval
pure_eval: pure_eval
Expand Down
Loading