You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/search/tutorial-rag-build-solution-minimize-storage.md
+4-6Lines changed: 4 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,9 +23,9 @@ In this tutorial, you modify the existing search index to use:
23
23
> - Scalar quantization
24
24
> - Reduced storage by opting out of vectors in search results
25
25
26
-
This tutorial reprises the search index created by the [indexing pipeline](tutorial-rag-build-solution-pipeline.md). All of these updates affect the existing content, requiring you to rerun the indexer. However, instead of deleting the search index, you create a second one so that you can compare reductions in vector index size after each incremental update.
26
+
This tutorial reprises the search index created by the [indexing pipeline](tutorial-rag-build-solution-pipeline.md). All of these updates affect the existing content, requiring you to rerun the indexer. However, instead of deleting the search index, you create a second one so that you can compare reductions in vector index size after adding the new capabilities.
27
27
28
-
Used together, these techniques can reduce vector storage by about half.
28
+
Altogether, the techniques illustrated in this tutorial can reduce vector storage by about half.
29
29
30
30
The following screenshot compares the [first index](tutorial-rag-build-solution-pipeline.md) from a previous tutorial to the index built in this one.
31
31
@@ -43,8 +43,6 @@ You should also have the following objects:
43
43
44
44
- py-rag-tutorial-ss (skillset)
45
45
46
-
- py-rag-tutorial-idxr (indexer)
47
-
48
46
## Download the sample
49
47
50
48
[Download a Jupyter notebook](https://github.com/Azure-Samples/azure-search-python-samples/blob/main/Tutorial-RAG/Tutorial-rag.ipynb) from GitHub to send the requests to Azure AI Search. For more information, see [Downloading files from GitHub](https://docs.github.com/get-started/start-your-journey/downloading-files-from-github).
@@ -66,7 +64,7 @@ All of these capabilities are specified in a search index. After you load the in
66
64
1. Use the following definition for the new index. The difference between this schema and the previous schema updates in [Maximize relevance](tutorial-rag-build-solution-maximize-relevance.md) are new classes for scalar quantization and a new compressions section, a new data type (`Collection(Edm.Half)`) for the text_vector field, and a new property `stored` set to false.
67
65
68
66
```python
69
-
from azure.identity import DefaultAzureCredential
67
+
from azure.identity import DefaultAzureCredential
70
68
from azure.identity import get_bearer_token_provider
71
69
from azure.search.documents.indexes import SearchIndexClient
72
70
from azure.search.documents.indexes.models import (
@@ -330,7 +328,7 @@ As a final step, switch to the Azure portal to compare the vector storage requir
330
328
331
329
The index created in this tutorial uses half-precision floating-point numbers (float16) for the text vectors. This reduces the storage requirements for the vectors by half compared to the previous index that used single-precision floating-point numbers (float32). Scalar compression and the omission of one set of the vectors account for the remaining storage savings. For more information about reducing vector size, see [Choose an approach for optimizing vector storage and processing](vector-search-how-to-configure-compression-storage.md).
332
330
333
-
Consider revisiting the queries from the previous tutorials so that you can compare query speed and utility. You should expect some variation in LLM output whenever you repeat a query, but in general the storage-saving techniques you implemented shouldn't degrade the quality of your search results.
331
+
Consider revisiting the [queries from the previous tutorial](tutorial-rag-build-solution-query.md) so that you can compare query speed and utility. You should expect some variation in LLM output whenever you repeat a query, but in general the storage-saving techniques you implemented shouldn't degrade the quality of your search results.
0 commit comments