Skip to content

Commit 73b9049

Browse files
committed
Black formatting
1 parent 3639c0f commit 73b9049

File tree

3 files changed

+69
-77
lines changed

3 files changed

+69
-77
lines changed

docs/langchain-retrieval-agent.ipynb

Lines changed: 30 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -354,7 +354,7 @@
354354
],
355355
"source": [
356356
"# we drop sparse_values as they are not needed for this example\n",
357-
"dataset.documents.drop(['sparse_values', 'blob'], axis=1, inplace=True)\n",
357+
"dataset.documents.drop([\"sparse_values\", \"blob\"], axis=1, inplace=True)\n",
358358
"\n",
359359
"dataset.head()"
360360
]
@@ -369,7 +369,7 @@
369369
"\n",
370370
"print(\"Here are some example topics in our Knowledge Base:\\n\")\n",
371371
"for r in dataset.documents.iloc[:].to_dict(orient=\"records\"):\n",
372-
" topics.add(r['metadata']['title'])\n",
372+
" topics.add(r[\"metadata\"][\"title\"])\n",
373373
"\n",
374374
"for topic in sorted(topics)[50:75]:\n",
375375
" print(f\"- {topic}\")"
@@ -396,6 +396,7 @@
396396
"\n",
397397
"if not os.environ.get(\"PINECONE_API_KEY\"):\n",
398398
" from pinecone_notebooks.colab import Authenticate\n",
399+
"\n",
399400
" Authenticate()"
400401
]
401402
},
@@ -464,18 +465,15 @@
464465
"source": [
465466
"from pinecone import ServerlessSpec\n",
466467
"\n",
467-
"index_name = 'langchain-retrieval-agent-fast'\n",
468+
"index_name = \"langchain-retrieval-agent-fast\"\n",
468469
"\n",
469470
"if not pc.has_index(name=index_name):\n",
470471
" # Create a new index\n",
471472
" pc.create_index(\n",
472473
" name=index_name,\n",
473474
" dimension=1536, # dimensionality of text-embedding-ada-002\n",
474-
" metric='dotproduct',\n",
475-
" spec=ServerlessSpec(\n",
476-
" cloud='aws',\n",
477-
" region='us-east-1'\n",
478-
" )\n",
475+
" metric=\"dotproduct\",\n",
476+
" spec=ServerlessSpec(cloud=\"aws\", region=\"us-east-1\"),\n",
479477
" )\n",
480478
"\n",
481479
"pc.describe_index(name=index_name)"
@@ -651,12 +649,9 @@
651649
"source": [
652650
"from langchain_openai import OpenAIEmbeddings\n",
653651
"\n",
654-
"openai_api_key = os.environ.get('OPENAI_API_KEY') or 'OPENAI_API_KEY'\n",
652+
"openai_api_key = os.environ.get(\"OPENAI_API_KEY\") or \"OPENAI_API_KEY\"\n",
655653
"\n",
656-
"embed = OpenAIEmbeddings(\n",
657-
" model='text-embedding-ada-002',\n",
658-
" openai_api_key=openai_api_key\n",
659-
")"
654+
"embed = OpenAIEmbeddings(model=\"text-embedding-ada-002\", openai_api_key=openai_api_key)"
660655
]
661656
},
662657
{
@@ -670,9 +665,7 @@
670665
"from langchain_pinecone import PineconeVectorStore\n",
671666
"\n",
672667
"pinecone_vectorstore = PineconeVectorStore(\n",
673-
" index_name=index_name, \n",
674-
" embedding=embed, \n",
675-
" text_key=\"text\"\n",
668+
" index_name=index_name, embedding=embed, text_key=\"text\"\n",
676669
")"
677670
]
678671
},
@@ -759,11 +752,12 @@
759752
"source": [
760753
"from pprint import pprint\n",
761754
"\n",
762-
"query = \"When was the college of engineering in the University of Notre Dame established?\"\n",
755+
"query = (\n",
756+
" \"When was the college of engineering in the University of Notre Dame established?\"\n",
757+
")\n",
763758
"\n",
764759
"documents = pinecone_vectorstore.similarity_search(\n",
765-
" query=query,\n",
766-
" k=3 # return 3 most relevant docs\n",
760+
" query=query, k=3 # return 3 most relevant docs\n",
767761
")\n",
768762
"\n",
769763
"for doc in documents:\n",
@@ -815,9 +809,7 @@
815809
"\n",
816810
"# Chat completion LLM\n",
817811
"llm = ChatOpenAI(\n",
818-
" openai_api_key=openai_api_key,\n",
819-
" model_name='gpt-3.5-turbo',\n",
820-
" temperature=0.0\n",
812+
" openai_api_key=openai_api_key, model_name=\"gpt-3.5-turbo\", temperature=0.0\n",
821813
")"
822814
]
823815
},
@@ -839,17 +831,19 @@
839831
"from langchain_core.runnables import RunnablePassthrough\n",
840832
"\n",
841833
"# Based on the RAG template from https://smith.langchain.com/hub/rlm/rag-prompt\n",
842-
"template=(\n",
834+
"template = (\n",
843835
" \"You are an assistant for question-answering tasks. Use the following pieces of retrieved context to answer the question. If you don't know the answer, just say that you don't know. Use three sentences maximum and keep the answer concise.\"\n",
844836
" \"Question: {question}\"\n",
845837
" \"Context: {context}\"\n",
846838
" \"Answer:\"\n",
847839
")\n",
848840
"prompt = PromptTemplate(input_variables=[\"question\", \"context\"], template=template)\n",
849841
"\n",
842+
"\n",
850843
"def format_docs(docs):\n",
851844
" return \"\\n\\n\".join(doc.page_content for doc in docs)\n",
852845
"\n",
846+
"\n",
853847
"# Retrieval Question-Answer chain\n",
854848
"qa_chain = (\n",
855849
" {\n",
@@ -896,7 +890,9 @@
896890
}
897891
],
898892
"source": [
899-
"qa_chain.invoke(\"When was the college of engineering in the University of Notre Dame established?\")"
893+
"qa_chain.invoke(\n",
894+
" \"When was the college of engineering in the University of Notre Dame established?\"\n",
895+
")"
900896
]
901897
},
902898
{
@@ -920,11 +916,11 @@
920916
"outputs": [],
921917
"source": [
922918
"knowledge_base_tool = qa_chain.as_tool(\n",
923-
" name='knowledge-base',\n",
924-
" description=(\n",
925-
" 'use this tool when answering general knowledge queries to get '\n",
926-
" 'more information about the topic'\n",
927-
" )\n",
919+
" name=\"knowledge-base\",\n",
920+
" description=(\n",
921+
" \"use this tool when answering general knowledge queries to get \"\n",
922+
" \"more information about the topic\"\n",
923+
" ),\n",
928924
")"
929925
]
930926
},
@@ -966,9 +962,11 @@
966962
"from langgraph.graph import StateGraph\n",
967963
"from langgraph.graph.message import add_messages\n",
968964
"\n",
965+
"\n",
969966
"class State(TypedDict):\n",
970967
" messages: Annotated[list, add_messages]\n",
971968
"\n",
969+
"\n",
972970
"graph_builder = StateGraph(State)"
973971
]
974972
},
@@ -1001,9 +999,11 @@
1001999
"tools = [knowledge_base_tool]\n",
10021000
"llm_with_tools = llm.bind_tools(tools)\n",
10031001
"\n",
1002+
"\n",
10041003
"def chatbot(state: State):\n",
10051004
" return {\"messages\": [llm_with_tools.invoke(state[\"messages\"])]}\n",
10061005
"\n",
1006+
"\n",
10071007
"graph_builder.add_node(\"chatbot\", chatbot)\n",
10081008
"\n",
10091009
"tool_node = ToolNode(tools=tools)\n",
@@ -1054,7 +1054,7 @@
10541054
"source": [
10551055
"def agent(user_message):\n",
10561056
" config = {\"configurable\": {\"thread_id\": \"1\"}}\n",
1057-
" \n",
1057+
"\n",
10581058
" # The config is the **second positional argument** to stream() or invoke()!\n",
10591059
" events = graph.stream(\n",
10601060
" {\"messages\": [{\"role\": \"user\", \"content\": user_message}]},\n",

docs/langchain-retrieval-augmentation.ipynb

Lines changed: 19 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -180,11 +180,11 @@
180180
"source": [
181181
"from pinecone_datasets import load_dataset\n",
182182
"\n",
183-
"dataset = load_dataset('wikipedia-simple-text-embedding-ada-002-50K')\n",
183+
"dataset = load_dataset(\"wikipedia-simple-text-embedding-ada-002-50K\")\n",
184184
"\n",
185185
"# We drop sparse_values and blob keys as they are not needed for this example\n",
186-
"dataset.documents.drop(['sparse_values'], axis=1, inplace=True)\n",
187-
"dataset.documents.drop(['blob'], axis=1, inplace=True)\n",
186+
"dataset.documents.drop([\"sparse_values\"], axis=1, inplace=True)\n",
187+
"dataset.documents.drop([\"blob\"], axis=1, inplace=True)\n",
188188
"\n",
189189
"dataset.head()"
190190
]
@@ -220,6 +220,7 @@
220220
"\n",
221221
"if not os.environ.get(\"PINECONE_API_KEY\"):\n",
222222
" from pinecone_notebooks.colab import Authenticate\n",
223+
"\n",
223224
" Authenticate()"
224225
]
225226
},
@@ -301,17 +302,14 @@
301302
"source": [
302303
"from pinecone import ServerlessSpec\n",
303304
"\n",
304-
"index_name = 'langchain-retrieval-augmentation-fast'\n",
305+
"index_name = \"langchain-retrieval-augmentation-fast\"\n",
305306
"\n",
306307
"if not pc.has_index(name=index_name):\n",
307308
" pc.create_index(\n",
308309
" name=index_name,\n",
309310
" dimension=1536, # dimensionality of text-embedding-ada-002\n",
310-
" metric='dotproduct',\n",
311-
" spec=ServerlessSpec(\n",
312-
" cloud='aws',\n",
313-
" region='us-east-1'\n",
314-
" )\n",
311+
" metric=\"dotproduct\",\n",
312+
" spec=ServerlessSpec(cloud=\"aws\", region=\"us-east-1\"),\n",
315313
" )\n",
316314
"\n",
317315
"pc.describe_index(name=index_name)"
@@ -420,8 +418,10 @@
420418
"\n",
421419
"batch_size = 100\n",
422420
"\n",
423-
"for start in tqdm(range(0, len(dataset.documents), batch_size), \"Upserting records batch\"):\n",
424-
" batch = dataset.documents.iloc[start:start + batch_size].to_dict(orient=\"records\")\n",
421+
"for start in tqdm(\n",
422+
" range(0, len(dataset.documents), batch_size), \"Upserting records batch\"\n",
423+
"):\n",
424+
" batch = dataset.documents.iloc[start : start + batch_size].to_dict(orient=\"records\")\n",
425425
" index.upsert(vectors=batch)"
426426
]
427427
},
@@ -487,14 +487,11 @@
487487
"from langchain_openai import OpenAIEmbeddings\n",
488488
"\n",
489489
"# Get openai api key from platform.openai.com\n",
490-
"OPENAI_API_KEY = os.getenv('OPENAI_API_KEY') or 'OPENAI_API_KEY'\n",
490+
"OPENAI_API_KEY = os.getenv(\"OPENAI_API_KEY\") or \"OPENAI_API_KEY\"\n",
491491
"\n",
492-
"model_name = 'text-embedding-ada-002'\n",
492+
"model_name = \"text-embedding-ada-002\"\n",
493493
"\n",
494-
"embed = OpenAIEmbeddings(\n",
495-
" model=model_name,\n",
496-
" openai_api_key=OPENAI_API_KEY\n",
497-
")"
494+
"embed = OpenAIEmbeddings(model=model_name, openai_api_key=OPENAI_API_KEY)"
498495
]
499496
},
500497
{
@@ -518,9 +515,7 @@
518515
"from langchain_pinecone import PineconeVectorStore\n",
519516
"\n",
520517
"pinecone_vectorstore = PineconeVectorStore(\n",
521-
" index_name=index_name, \n",
522-
" embedding=embed, \n",
523-
" text_key=\"text\"\n",
518+
" index_name=index_name, embedding=embed, text_key=\"text\"\n",
524519
")"
525520
]
526521
},
@@ -670,7 +665,7 @@
670665
"\n",
671666
"documents = pinecone_vectorstore.similarity_search(\n",
672667
" query=\"Who was Benito Mussolini?\", # our search query\n",
673-
" k=3 # return 3 most relevant docs\n",
668+
" k=3, # return 3 most relevant docs\n",
674669
")\n",
675670
"\n",
676671
"for doc in documents:\n",
@@ -707,15 +702,11 @@
707702
"\n",
708703
"# Chat Completion LLM\n",
709704
"llm = ChatOpenAI(\n",
710-
" openai_api_key=OPENAI_API_KEY,\n",
711-
" model_name='gpt-4.5-preview',\n",
712-
" temperature=0.0\n",
705+
" openai_api_key=OPENAI_API_KEY, model_name=\"gpt-4.5-preview\", temperature=0.0\n",
713706
")\n",
714707
"\n",
715708
"qa = RetrievalQA.from_chain_type(\n",
716-
" llm=llm,\n",
717-
" chain_type=\"stuff\",\n",
718-
" retriever=pinecone_vectorstore.as_retriever()\n",
709+
" llm=llm, chain_type=\"stuff\", retriever=pinecone_vectorstore.as_retriever()\n",
719710
")"
720711
]
721712
},
@@ -771,7 +762,7 @@
771762
" llm=llm,\n",
772763
" chain_type=\"stuff\",\n",
773764
" retriever=pinecone_vectorstore.as_retriever(),\n",
774-
" return_source_documents=True\n",
765+
" return_source_documents=True,\n",
775766
")"
776767
]
777768
},

0 commit comments

Comments
 (0)