-
Notifications
You must be signed in to change notification settings - Fork 4.5k
complete implementation of open ai text embedding with test #new #34700
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 24 commits
0c43846
5b60a99
7771ef8
e25e5ee
0760d5c
ddb6907
b327620
ccdf259
99dc050
002c98f
86d2e03
9b1d20a
f15dfb0
c3dc96e
cf58224
3ba997b
7f3df7c
fbafd11
f73e144
e7a5533
7cd37a3
381d835
433958d
fb48432
53fade7
afc22e9
6247c07
34962b1
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,204 @@ | ||
| # Licensed to the Apache Software Foundation (ASF) under one or more | ||
| # contributor license agreements. See the NOTICE file distributed with | ||
| # this work for additional information regarding copyright ownership. | ||
| # The ASF licenses this file to You under the Apache License, Version 2.0 | ||
| # (the "License"); you may not use this file except in compliance with | ||
| # the License. You may obtain a copy of the License at | ||
| # | ||
| # http://www.apache.org/licenses/LICENSE-2.0 | ||
| # | ||
| # Unless required by applicable law or agreed to in writing, software | ||
| # distributed under the License is distributed on an "AS IS" BASIS, | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| # See the License for the specific language governing permissions and | ||
| # limitations under the License. | ||
|
|
||
| import logging | ||
| from collections.abc import Iterable | ||
| from collections.abc import Sequence | ||
| from typing import Any | ||
| from typing import Optional | ||
| from typing import TypeVar | ||
| from typing import Union | ||
| from typing import cast | ||
|
|
||
| import apache_beam as beam | ||
| import openai | ||
| from apache_beam.ml.inference.base import RemoteModelHandler | ||
| from apache_beam.ml.inference.base import RunInference | ||
| from apache_beam.ml.transforms.base import EmbeddingsManager | ||
| from apache_beam.ml.transforms.base import _TextEmbeddingHandler | ||
| from apache_beam.pvalue import PCollection | ||
| from apache_beam.pvalue import Row | ||
| from openai import APIError | ||
| from openai import RateLimitError | ||
|
|
||
| __all__ = ["OpenAITextEmbeddings"] | ||
|
|
||
| # Define a type variable for the output | ||
| MLTransformOutputT = TypeVar('MLTransformOutputT') | ||
|
|
||
| _BATCH_SIZE = 20 | ||
|
|
||
| LOGGER = logging.getLogger("OpenAIEmbeddings") | ||
|
|
||
|
|
||
| def _retry_on_appropriate_openai_error(exception): # pylint: disable=line-too-long | ||
| """ | ||
| Retry filter that returns True for rate limit (429) or server (5xx) errors. | ||
|
|
||
| Args: | ||
| exception: the returned exception encountered during the request/response | ||
| loop. | ||
|
|
||
| Returns: | ||
| boolean indication whether or not the exception is a Server Error (5xx) or | ||
| a RateLimitError (429) error. | ||
| """ | ||
| return isinstance(exception, (RateLimitError, APIError)) | ||
|
|
||
|
|
||
| class _OpenAITextEmbeddingHandler(RemoteModelHandler): | ||
| """ | ||
| Note: Intended for internal use and guarantees no backwards compatibility. | ||
| """ | ||
| def __init__( | ||
| self, | ||
| model_name: str, | ||
| api_key: Optional[str] = None, | ||
| organization: Optional[str] = None, | ||
| dimensions: Optional[int] = None, | ||
| user: Optional[str] = None, | ||
| batch_size: Optional[int] = None, | ||
| ): | ||
| super().__init__( | ||
| namespace="OpenAITextEmbeddings", | ||
| num_retries=5, | ||
| throttle_delay_secs=5, | ||
| retry_filter=_retry_on_appropriate_openai_error) | ||
| self.model_name = model_name | ||
| self.api_key = api_key | ||
| self.organization = organization | ||
| self.dimensions = dimensions | ||
| self.user = user | ||
| self.batch_size = batch_size or _BATCH_SIZE | ||
|
|
||
| def create_client(self): | ||
| """Creates and returns an OpenAI client.""" | ||
| if self.api_key: | ||
| client = openai.OpenAI( | ||
| api_key=self.api_key, | ||
| organization=self.organization, | ||
| ) | ||
| else: | ||
| client = openai.OpenAI(organization=self.organization) | ||
|
|
||
| return client | ||
|
|
||
| def request( | ||
| self, | ||
| batch: Sequence[str], | ||
| model: Any, | ||
| inference_args: Optional[dict[str, Any]] = None, | ||
| ) -> Iterable: | ||
| """Makes a request to OpenAI embedding API and returns embeddings.""" | ||
| # Process in smaller batches if needed | ||
| if len(batch) > self.batch_size: | ||
| embeddings = [] | ||
| for i in range(0, len(batch), self.batch_size): | ||
| text_batch = batch[i:i + self.batch_size] | ||
| # Use request() recursively for each smaller batch | ||
| embeddings_batch = self.request(text_batch, model, inference_args) | ||
| embeddings.extend(embeddings_batch) | ||
| return embeddings | ||
aditya0yadav marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
|
|
||
| # Prepare arguments for the API call | ||
| kwargs = { | ||
| "model": self.model_name, | ||
| "input": batch, | ||
| } | ||
| if self.dimensions: | ||
| kwargs["dimensions"] = cast(Any, self.dimensions) | ||
aditya0yadav marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
| if self.user: | ||
| kwargs["user"] = self.user | ||
|
|
||
| try: | ||
| # Make the API call | ||
| response = model.embeddings.create(**kwargs) | ||
| return [item.embedding for item in response.data] | ||
| except RateLimitError as e: | ||
| LOGGER.warning("Request was rate limited by OpenAI API: %s", e) | ||
| raise | ||
| except Exception as e: | ||
| LOGGER.error("Unexpected exception raised as part of request: %s", e) | ||
| raise | ||
aditya0yadav marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
|
|
||
| def batch_elements_kwargs(self) -> dict[str, Any]: | ||
| """Returns kwargs suitable for beam.BatchElements with appropriate batch size.""" # pylint: disable=line-too-long | ||
| return {'max_batch_size': self.batch_size} | ||
|
|
||
| def __repr__(self): | ||
| return 'OpenAITextEmbeddings' | ||
|
|
||
|
|
||
| class OpenAITextEmbeddings(EmbeddingsManager): | ||
| """ | ||
| A PTransform that uses OpenAI's API to generate embeddings from text inputs. | ||
|
|
||
| Example Usage:: | ||
|
|
||
| with pipeline as p: # pylint: disable=line-too-long | ||
| text = p | "Create texts" >> beam.Create([{"text": "Hello world"}, | ||
| {"text": "Beam ML"}]) | ||
| embeddings = text | OpenAITextEmbeddings( | ||
| model_name="text-embedding-3-small", | ||
| columns=["embedding_col"], | ||
| api_key=api_key | ||
| ) | ||
| """ | ||
| @beam.typehints.with_output_types(PCollection[Union[MLTransformOutputT, Row]]) # pylint: disable=line-too-long | ||
| def __init__( | ||
| self, | ||
| model_name: str, | ||
| columns: list[str], | ||
| api_key: Optional[str] = None, | ||
| organization: Optional[str] = None, | ||
| dimensions: Optional[int] = None, | ||
| user: Optional[str] = None, | ||
| batch_size: Optional[int] = None, | ||
|
||
| **kwargs): | ||
| """ | ||
| Embedding Config for OpenAI Text Embedding models. | ||
| Text Embeddings are generated for a batch of text using the OpenAI API. | ||
|
|
||
| Args: | ||
| model_name: Name of the OpenAI embedding model | ||
| columns: The columns where the embeddings will be stored in the output | ||
| api_key: OpenAI API key | ||
| organization: OpenAI organization ID | ||
| dimensions: Specific embedding dimensions to use (if model supports it) | ||
| user: End-user identifier for tracking and rate limit calculations | ||
| batch_size: Maximum batch size for requests to OpenAI API (default: 20) | ||
| """ | ||
| self.model_name = model_name | ||
| self.api_key = api_key | ||
| self.organization = organization | ||
| self.dimensions = dimensions | ||
| self.user = user | ||
| self.batch_size = batch_size | ||
| super().__init__(columns=columns, **kwargs) | ||
|
|
||
| def get_model_handler(self) -> RemoteModelHandler: | ||
| return _OpenAITextEmbeddingHandler( | ||
| model_name=self.model_name, | ||
| api_key=self.api_key, | ||
| organization=self.organization, | ||
| dimensions=self.dimensions, | ||
| user=self.user, | ||
| batch_size=self.batch_size, | ||
| ) | ||
|
|
||
| def get_ptransform_for_processing(self, **kwargs) -> beam.PTransform: | ||
| return RunInference( | ||
| model_handler=_TextEmbeddingHandler(self), | ||
| inference_args=self.inference_args) | ||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Need an import for these exceptions
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
because of the unknown reason some of the import is missing i just add those import
it will be shown in a upcoming update commit