You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: scenarios/AksKaito/README.md
-10Lines changed: 0 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,16 +9,6 @@ ms.author: schaffererin
9
9
10
10
---
11
11
12
-
## Quickstart: Create a Linux virtual machine with the Azure CLI on Azure
13
-
14
-
**Applies to:**:heavy_check_mark: Linux VMs
15
-
16
-
[](https://go.microsoft.com/fwlink/?linkid=2262692)
17
-
18
-
This quickstart shows you how to use the Azure CLI to deploy a Linux virtual machine (VM) in Azure. The Azure CLI is used to create and manage Azure resources via either the command line or scripts.
19
-
20
-
If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/free/?WT.mc_id=A261C142F) before you begin.
21
-
22
12
## Deploy an AI model on Azure Kubernetes Service (AKS) with the AI toolchain operator (preview)
23
13
24
14
The AI toolchain operator (KAITO) is a managed add-on for AKS that simplifies the experience of running OSS AI models on your AKS clusters. The AI toolchain operator automatically provisions the necessary GPU nodes and sets up the associated inference server as an endpoint server to your AI models. Using this add-on reduces your onboarding time and enables you to focus on AI model usage and development rather than infrastructure setup.
In this doc, we go over how to host the infrastructure required to run a basic LLM model with RAG capabilities on Azure.
14
+
We first set up a Postgres database capable of storing vector embeddings for documents/knowledge files that we want to use to
15
+
augment our queries. We then create an Azure OpenAI deployment capable of generating embeddings and answering questions using the latest 'gpt-4-turbo' model.
16
+
We then use a python script to fill our postgres database with embeddings from a sample "knowledge.txt" file containing information about an imaginary
17
+
resource called 'Zytonium'. Once the database is filled with those embeddings, we use the same python script to answer any
18
+
questions we have about 'Zytonium'. The script will search the database for relevant information for our query using an embeddings search and
19
+
then augment our query with that relevant information before being sent our LLM to answer.
0 commit comments