Skip to content

Commit e82b37b

Browse files
authored
Complete the LLM deployment (#139)
* make deployed version work * set 4o as the new default * formatting
1 parent d259599 commit e82b37b

17 files changed

+43
-51
lines changed

llm-complete-guide/constants.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
RATE_LIMIT = 5 # Maximum number of requests per second
2828

2929
# LLM Utils constants
30-
OPENAI_MODEL = "gpt-3.5-turbo"
30+
OPENAI_MODEL = "gpt-4o"
3131
EMBEDDINGS_MODEL = "sentence-transformers/all-MiniLM-L12-v2"
3232
MODEL_NAME_MAP = {
3333
"gpt4": "gpt-4",

llm-complete-guide/most_basic_eval.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,12 +14,10 @@
1414
# See the License for the specific language governing permissions and
1515
# limitations under the License.
1616

17-
import os
1817
import re
1918
import string
2019

2120
from openai import OpenAI
22-
2321
from utils.openai_utils import get_openai_api_key
2422

2523

llm-complete-guide/most_basic_rag_pipeline.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,12 +15,10 @@
1515
# limitations under the License.
1616
#
1717

18-
import os
1918
import re
2019
import string
2120

2221
from openai import OpenAI
23-
2422
from utils.openai_utils import get_openai_api_key
2523

2624

llm-complete-guide/notebooks/argilla_embeddings.ipynb

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -68,10 +68,10 @@
6868
"source": [
6969
"import os\n",
7070
"\n",
71-
"from distilabel.steps.tasks import GenerateSentencePair\n",
72-
"from distilabel.llms import OpenAILLM, OllamaLLM\n",
73-
"from distilabel.steps import LoadDataFromHub\n",
71+
"from distilabel.llms import OpenAILLM\n",
7472
"from distilabel.pipeline import Pipeline\n",
73+
"from distilabel.steps import LoadDataFromHub\n",
74+
"from distilabel.steps.tasks import GenerateSentencePair\n",
7575
"\n",
7676
"# TODO: I think we might optimize this a bit more.\n",
7777
"\n",
@@ -254,8 +254,8 @@
254254
"metadata": {},
255255
"outputs": [],
256256
"source": [
257-
"from sentence_transformers import SentenceTransformer\n",
258257
"import torch\n",
258+
"from sentence_transformers import SentenceTransformer\n",
259259
"\n",
260260
"model_id = \"Snowflake/snowflake-arctic-embed-m\" # Hugging Face model ID\n",
261261
"\n",
@@ -508,8 +508,8 @@
508508
"outputs": [],
509509
"source": [
510510
"import argilla as rg\n",
511-
"from zenml.client import Client\n",
512511
"from argilla._exceptions import ConflictError\n",
512+
"from zenml.client import Client\n",
513513
"\n",
514514
"zenml_client = Client()\n",
515515
"api_key = zenml_client.get_secret(\"argilla_hf\").secret_values[\"api_key\"]\n",
@@ -654,13 +654,13 @@
654654
"outputs": [],
655655
"source": [
656656
"import torch\n",
657+
"from datasets import concatenate_datasets, load_dataset\n",
657658
"from sentence_transformers import SentenceTransformer\n",
658659
"from sentence_transformers.evaluation import (\n",
659660
" InformationRetrievalEvaluator,\n",
660661
" SequentialEvaluator,\n",
661662
")\n",
662663
"from sentence_transformers.util import cos_sim\n",
663-
"from datasets import load_dataset, concatenate_datasets\n",
664664
"\n",
665665
"model_id = \"Snowflake/snowflake-arctic-embed-m\" # Hugging Face model ID\n",
666666
"matryoshka_dimensions = [384, 256, 128, 64] # Important: large to small\n",
@@ -752,8 +752,8 @@
752752
"outputs": [],
753753
"source": [
754754
"from sentence_transformers import (\n",
755-
" SentenceTransformerModelCardData,\n",
756755
" SentenceTransformer,\n",
756+
" SentenceTransformerModelCardData,\n",
757757
")\n",
758758
"\n",
759759
"model_id = \"Snowflake/snowflake-arctic-embed-m\"\n",

llm-complete-guide/notebooks/finetune_embeddings.ipynb

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,6 @@
6464
"outputs": [],
6565
"source": [
6666
"from datasets import Dataset, DatasetDict\n",
67-
"from huggingface_hub import create_repo\n",
6867
"from zenml.client import Client\n",
6968
"\n",
7069
"# Assuming you have a Dataset object named 'dataset'\n",
@@ -174,9 +173,8 @@
174173
"metadata": {},
175174
"outputs": [],
176175
"source": [
177-
"from torch.utils.data import DataLoader\n",
178176
"from sentence_transformers import losses\n",
179-
"\n",
177+
"from torch.utils.data import DataLoader\n",
180178
"\n",
181179
"train_dataloaderB = DataLoader(train_examplesB, shuffle=True, batch_size=64)\n",
182180
"train_lossB = losses.MultipleNegativesRankingLoss(model=modelB)\n",
@@ -218,8 +216,8 @@
218216
}
219217
],
220218
"source": [
221-
"from torch.utils.data import Dataset, DataLoader\n",
222219
"from sentence_transformers import InputExample\n",
220+
"from torch.utils.data import DataLoader, Dataset\n",
223221
"\n",
224222
"\n",
225223
"class InputExampleDataset(Dataset):\n",

llm-complete-guide/notebooks/visualise_embeddings.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,8 @@
1616
],
1717
"source": [
1818
"import matplotlib.pyplot as plt\n",
19-
"from sklearn.manifold import TSNE\n",
2019
"import umap\n",
20+
"from sklearn.manifold import TSNE\n",
2121
"from zenml.client import Client\n",
2222
"\n",
2323
"artifact = Client().get_artifact_version(\n",
@@ -32,8 +32,8 @@
3232
"metadata": {},
3333
"outputs": [],
3434
"source": [
35-
"import numpy as np\n",
3635
"import matplotlib.pyplot as plt\n",
36+
"import numpy as np\n",
3737
"from matplotlib.colors import ListedColormap\n",
3838
"\n",
3939
"embeddings = np.array([doc.embedding for doc in documents])\n",

llm-complete-guide/pipelines/llm_basic_rag.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
)
2323
from steps.url_scraper import url_scraper
2424
from steps.web_url_loader import web_url_loader
25-
from zenml import pipeline, Model
25+
from zenml import pipeline
2626

2727

2828
@pipeline

llm-complete-guide/pipelines/rag_deployment.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
1-
from zenml import pipeline
21
from steps.rag_deployment import gradio_rag_deployment
2+
from zenml import pipeline
33

44

55
@pipeline(enable_cache=False)

llm-complete-guide/steps/distilabel_generate_queries.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,6 @@
1212
# or implied. See the License for the specific language governing
1313
# permissions and limitations under the License.
1414

15-
import os
1615
from typing import Annotated, Tuple
1716

1817
from constants import (
@@ -22,12 +21,11 @@
2221
)
2322
from datasets import Dataset
2423
from distilabel.llms import OpenAILLM
24+
from distilabel.pipeline import Pipeline
2525
from distilabel.steps import LoadDataFromHub
2626
from distilabel.steps.tasks import GenerateSentencePair
27-
from distilabel.pipeline import Pipeline
28-
from zenml import step
29-
3027
from utils.openai_utils import get_openai_api_key
28+
from zenml import step
3129

3230
synthetic_generation_context = """
3331
The text is a chunk from technical documentation of ZenML.

llm-complete-guide/steps/eval_e2e.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,9 +23,8 @@
2323
from pydantic import BaseModel, conint
2424
from structures import TestResult
2525
from utils.llm_utils import process_input_with_retrieval
26-
from zenml import step
27-
2826
from utils.openai_utils import get_openai_api_key
27+
from zenml import step
2928

3029
logging.getLogger().setLevel(logging.WARNING)
3130

0 commit comments

Comments
 (0)