Skip to content

Commit 0ea8800

Browse files
fix(docs): correct multiple typos in pdf_parsing_for_semantic_retrieval_systems.ipynb
1 parent 61a3666 commit 0ea8800

File tree

1 file changed

+8
-8
lines changed

1 file changed

+8
-8
lines changed

site/en/docs/pdf_parsing_for_semantic_retrieval_systems.ipynb

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -742,7 +742,7 @@
742742
" questions: list[str],\n",
743743
" links: list[str],\n",
744744
" ) -> dict[str, Any]:\n",
745-
" \"\"\"Structred data extraction from image analysis.\"\"\"\n",
745+
" \"\"\"Structured data extraction from image analysis.\"\"\"\n",
746746
" return {\n",
747747
" 'title': title,\n",
748748
" 'key_words': key_words,\n",
@@ -774,7 +774,7 @@
774774
" function_declarations=[glm.FunctionDeclaration(\n",
775775
" name=\"structured_data_extraction\",\n",
776776
" description=textwrap.dedent(\"\"\"\\\n",
777-
" Structred data extraction from image analysis.\n",
777+
" Structured data extraction from image analysis.\n",
778778
" \"\"\"),\n",
779779
" parameters=glm.Schema(\n",
780780
" type=glm.Type.OBJECT,\n",
@@ -815,20 +815,20 @@
815815
" \"\"\"Extracts metadata from the image provided and returns it in a structured dict.\"\"\"\n",
816816
" prompt = textwrap.dedent(f\"\"\"\n",
817817
" You are an expert image analyzer. Given an image of a PDF page, your job is to write the following for each and every image.\n",
818-
" 1. Generate key-words that matches the content from the image. (at most 10.)\n",
818+
" 1. Generate key-words that match the content from the image. (at most 10.)\n",
819819
" 2. Suggest a one-word title for the image.\n",
820820
" 3. Generate 1-2 short questions from the image.\n",
821821
" 4. Extract links that are present in the image.\n",
822822
"\n",
823823
" Your answer should follow the following format.\n",
824824
" ** 1. Key-words**\n",
825-
" [list of relevant key-words to descibe the content of the image]\n",
825+
" [list of relevant key-words to describe the content of the image]\n",
826826
"\n",
827827
" **2. Title**\n",
828828
" Suggest a one-word title based on the content here.\n",
829829
"\n",
830830
" **3. Questions**\n",
831-
" [lst of generated questions here...]\n",
831+
" [list of generated questions here...]\n",
832832
" ....\n",
833833
"\n",
834834
" **4. Links**\n",
@@ -952,7 +952,7 @@
952952
"id": "-1q_v21t2E94"
953953
},
954954
"source": [
955-
"Neat! The models were successfuly able to extract your custom metadata from the given information sources!"
955+
"Neat! The models were successfully able to extract your custom metadata from the given information sources!"
956956
]
957957
},
958958
{
@@ -1019,7 +1019,7 @@
10191019
" is_separator_regex=False,\n",
10201020
" )\n",
10211021
"\n",
1022-
" # iter through all PDF files.\n",
1022+
" # iterate through all PDF files.\n",
10231023
" for filename, file_bytes in pdfs.items():\n",
10241024
" print(f\"Extracting data from file: {filename}\")\n",
10251025
"\n",
@@ -1239,7 +1239,7 @@
12391239
"id": "vZl-A8EMVCZu"
12401240
},
12411241
"source": [
1242-
"`relevant_chunks` has chunks that matched our search results. Each chunk returned has a `chunk_relevance_score` and `chunk`. Where `chunk_relevance_score` deontes the degree to which the `user_query` is semantically similar to the contents from `chunk`."
1242+
"`relevant_chunks` has chunks that matched our search results. Each chunk returned has a `chunk_relevance_score` and `chunk`. Where `chunk_relevance_score` denotes the degree to which the `user_query` is semantically similar to the contents from `chunk`."
12431243
]
12441244
},
12451245
{

0 commit comments

Comments
 (0)