Skip to content

Commit d0e74ea

Browse files
committed
PR fixes
1 parent 5a6e48b commit d0e74ea

File tree

2 files changed

+10
-10
lines changed

2 files changed

+10
-10
lines changed

scenarios/AksKaito/README.md

Lines changed: 0 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -9,16 +9,6 @@ ms.author: schaffererin
99

1010
---
1111

12-
## Quickstart: Create a Linux virtual machine with the Azure CLI on Azure
13-
14-
**Applies to:** :heavy_check_mark: Linux VMs
15-
16-
[![Deploy to Azure](https://aka.ms/deploytoazurebutton)](https://go.microsoft.com/fwlink/?linkid=2262692)
17-
18-
This quickstart shows you how to use the Azure CLI to deploy a Linux virtual machine (VM) in Azure. The Azure CLI is used to create and manage Azure resources via either the command line or scripts.
19-
20-
If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/free/?WT.mc_id=A261C142F) before you begin.
21-
2212
## Deploy an AI model on Azure Kubernetes Service (AKS) with the AI toolchain operator (preview)
2313

2414
The AI toolchain operator (KAITO) is a managed add-on for AKS that simplifies the experience of running OSS AI models on your AKS clusters. The AI toolchain operator automatically provisions the necessary GPU nodes and sets up the associated inference server as an endpoint server to your AI models. Using this add-on reduces your onboarding time and enables you to focus on AI model usage and development rather than infrastructure setup.

scenarios/PostgresRagLlmDemo/README.md

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,16 @@ ms.author: ariaamini
88
ms.custom: innovation-engine, linux-related-content
99
---
1010

11+
## Introduction
12+
13+
In this doc, we go over how to host the infrastructure required to run a basic LLM model with RAG capabilities on Azure.
14+
We first set up a Postgres database capable of storing vector embeddings for documents/knowledge files that we want to use to
15+
augment our queries. We then create an Azure OpenAI deployment capable of generating embeddings and answering questions using the latest 'gpt-4-turbo' model.
16+
We then use a python script to fill our postgres database with embeddings from a sample "knowledge.txt" file containing information about an imaginary
17+
resource called 'Zytonium'. Once the database is filled with those embeddings, we use the same python script to answer any
18+
questions we have about 'Zytonium'. The script will search the database for relevant information for our query using an embeddings search and
19+
then augment our query with that relevant information before being sent our LLM to answer.
20+
1121
## Set up resource group
1222

1323
Set up a resource group with a random ID.

0 commit comments

Comments
 (0)