Skip to content

Commit 8cfabc3

Browse files
committed
fix lint
1 parent 01c87a5 commit 8cfabc3

File tree

7 files changed

+10
-28
lines changed

7 files changed

+10
-28
lines changed
Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,2 @@
11
[pycodestyle]
2-
count = True
3-
# E501 - max line length - will be left to the developer to determine if something is too long
4-
# W503 - line break before binary operator - is not PEP8 compliant
5-
ignore = E501, W503
6-
max-line-length = 200
7-
2+
max-line-length = 120

packages/ai-providers/server-ai-openai/src/ldai_openai/__init__.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,4 +3,3 @@
33
from ldai_openai.openai_provider import OpenAIProvider
44

55
__all__ = ['OpenAIProvider']
6-

packages/ai-providers/server-ai-openai/src/ldai_openai/openai_provider.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,13 @@
44
import os
55
from typing import Any, Dict, Iterable, List, Optional, cast
66

7-
from openai import AsyncOpenAI
8-
from openai.types.chat import ChatCompletionMessageParam
9-
107
from ldai import LDMessage
118
from ldai.models import AIConfigKind
129
from ldai.providers import AIProvider
1310
from ldai.providers.types import ChatResponse, LDAIMetrics, StructuredResponse
1411
from ldai.tracker import TokenUsage
12+
from openai import AsyncOpenAI
13+
from openai.types.chat import ChatCompletionMessageParam
1514

1615

1716
class OpenAIProvider(AIProvider):
@@ -250,4 +249,3 @@ def create_ai_metrics(openai_response: Any) -> LDAIMetrics:
250249
:return: LDAIMetrics with success status and token usage
251250
"""
252251
return OpenAIProvider.get_ai_metrics_from_response(openai_response)
253-
Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,2 @@
11
[pycodestyle]
2-
count = True
3-
# E501 - max line length - will be left to the developer to determine if something is too long
4-
# W503 - line break before binary operator - is not PEP8 compliant
5-
ignore = E501, W503
6-
max-line-length = 200
7-
2+
max-line-length = 120
Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
"""LaunchDarkly AI SDK Vercel Provider (Multi-Provider Support via LiteLLM)."""
22

3-
from ldai_vercel.vercel_provider import VercelProvider
43
from ldai_vercel.types import (
4+
ModelUsageTokens,
5+
StreamResponse,
6+
TextResponse,
57
VercelModelParameters,
8+
VercelProviderFunction,
69
VercelSDKConfig,
710
VercelSDKMapOptions,
8-
VercelProviderFunction,
9-
ModelUsageTokens,
10-
TextResponse,
11-
StreamResponse,
1211
)
12+
from ldai_vercel.vercel_provider import VercelProvider
1313

1414
__all__ = [
1515
'VercelProvider',
@@ -21,4 +21,3 @@
2121
'TextResponse',
2222
'StreamResponse',
2323
]
24-

packages/ai-providers/server-ai-vercel/src/ldai_vercel/types.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55

66
from ldai import LDMessage
77

8-
98
# Type alias for provider function
109
VercelProviderFunction = Callable[[str], Any]
1110

@@ -115,4 +114,3 @@ class StreamResponse:
115114
# Note: In async streaming, these would be resolved after the stream completes
116115
finish_reason: Optional[str] = None
117116
usage: Optional[ModelUsageTokens] = None
118-

packages/ai-providers/server-ai-vercel/src/ldai_vercel/vercel_provider.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,13 +4,12 @@
44
from typing import Any, Callable, Dict, List, Optional, Union
55

66
import litellm
7-
from litellm import acompletion
8-
97
from ldai import LDMessage
108
from ldai.models import AIConfigKind
119
from ldai.providers import AIProvider
1210
from ldai.providers.types import ChatResponse, LDAIMetrics, StructuredResponse
1311
from ldai.tracker import TokenUsage
12+
from litellm import acompletion
1413

1514
from ldai_vercel.types import (
1615
ModelUsageTokens,
@@ -391,4 +390,3 @@ def create_model_name(ai_config: AIConfigKind) -> str:
391390
if provider_name:
392391
return f'{provider_name}/{model_name}'
393392
return model_name
394-

0 commit comments

Comments
 (0)