Skip to content

Commit a917c29

Browse files
committed
Merge branch 'release-build-2025-release' of github.com:MicrosoftDocs/azure-ai-docs-pr into sdg-patches
2 parents c42d06c + 80d9211 commit a917c29

19 files changed

+156
-89
lines changed

articles/ai-foundry/concepts/models-featured.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -251,9 +251,7 @@ See [the Microsoft model collection in Azure AI Foundry portal](https://ai.azure
251251
Mistral AI offers two categories of models, namely:
252252

253253
- _Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, Mistral Medium 3 (25.05), and Ministral 3B models, and are available as serverless APIs with pay-as-you-go token-based billing.
254-
- _Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments.
255-
- _Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
256-
254+
- _Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as serverless APIs with pay-as-you-go token-based billing), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
257255

258256
| Model | Type | Capabilities |
259257
| ------ | ---- | --- |

articles/ai-foundry/foundry-local/concepts/foundry-local-architecture.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,8 @@ author: samuel100
1313

1414
# Foundry Local architecture
1515

16+
[!INCLUDE [foundry-local-preview](./../includes/foundry-local-preview.md)]
17+
1618
Foundry Local enables efficient, secure, and scalable AI model inference directly on your devices. This article explains the core components of Foundry Local and how they work together to deliver AI capabilities.
1719

1820
Key benefits of Foundry Local include:
@@ -37,7 +39,7 @@ The Foundry Local architecture consists of these main components:
3739

3840
The Foundry Local Service includes an OpenAI-compatible REST server that provides a standard interface for working with the inference engine. It's also possible to manage models over REST. Developers use this API to send requests, run models, and get results programmatically.
3941

40-
- **Endpoint**: The endpoint is *dynamically allocated* when the service starts. You can find the endpoint by running the `foundry service status` command. When using Foundry Local in your applications, we recommend using the SDK that automatically handles the endpoint for you. For more details on how to use the Foundry Local SDK, read the [Integrated inferencing SDKs with Foundry Local](../how-to/how-to-integrate-with-inference-sdks.md) article.
42+
- **Endpoint**: The endpoint is _dynamically allocated_ when the service starts. You can find the endpoint by running the `foundry service status` command. When using Foundry Local in your applications, we recommend using the SDK that automatically handles the endpoint for you. For more details on how to use the Foundry Local SDK, read the [Integrated inferencing SDKs with Foundry Local](../how-to/how-to-integrate-with-inference-sdks.md) article.
4143
- **Use Cases**:
4244
- Connect Foundry Local to your custom applications
4345
- Execute models through HTTP requests
@@ -109,9 +111,7 @@ The Foundry CLI is a powerful tool for managing models, the inference engine, an
109111
110112
#### Inferencing SDK integration
111113

112-
Foundry Local supports integration with various SDKs, such as the OpenAI SDK, enabling developers to use familiar programming interfaces to interact with the local inference engine.
113-
114-
- **Supported SDKs**: Python, JavaScript, C#, and more.
114+
Foundry Local supports integration with various SDKs in most languages, such as the OpenAI SDK, enabling developers to use familiar programming interfaces to interact with the local inference engine.
115115

116116
> [!TIP]
117117
> To learn more about integrating with inferencing SDKs, read [Integrate inferencing SDKs with Foundry Local](../how-to/how-to-integrate-with-inference-sdks.md).

articles/ai-foundry/foundry-local/get-started.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,8 @@ ms.custom: build-2025
1616

1717
# Get started with Foundry Local
1818

19+
[!INCLUDE [foundry-local-preview](./includes/foundry-local-preview.md)]
20+
1921
This guide walks you through setting up Foundry Local to run AI models on your device.
2022

2123
## Prerequisites

articles/ai-foundry/foundry-local/how-to/how-to-chat-application-with-open-web-ui.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,8 @@ ms.custom: build-2025
1616

1717
# Integrate Open Web UI with Foundry Local
1818

19+
[!INCLUDE [foundry-local-preview](./../includes/foundry-local-preview.md)]
20+
1921
This tutorial shows you how to create a chat application using Foundry Local and Open Web UI. When you finish, you have a working chat interface running entirely on your local device.
2022

2123
## Prerequisites

articles/ai-foundry/foundry-local/how-to/how-to-compile-hugging-face-models.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,8 @@ author: samuel100
1313

1414
# Compile Hugging Face models to run on Foundry Local
1515

16+
[!INCLUDE [foundry-local-preview](./../includes/foundry-local-preview.md)]
17+
1618
Foundry Local runs ONNX models on your device with high performance. While the model catalog offers _out-of-the-box_ precompiled options, you can use any model in the ONNX format.
1719

1820
To compile existing models in Safetensor or PyTorch format into the ONNX format, you can use [Olive](https://microsoft.github.io/Olive). Olive is a tool that optimizes models to ONNX format, making them suitable for deployment in Foundry Local. It uses techniques like _quantization_ and _graph optimization_ to improve performance.

articles/ai-foundry/foundry-local/how-to/how-to-integrate-with-inference-sdks.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,8 @@ author: samuel100
1414

1515
# Integrate inferencing SDKs with Foundry Local
1616

17+
[!INCLUDE [foundry-local-preview](./../includes/foundry-local-preview.md)]
18+
1719
Foundry Local integrates with various inferencing SDKs - such as OpenAI, Azure OpenAI, Langchain, etc. This guide shows you how to connect your applications to locally running AI models using popular SDKs.
1820

1921
## Prerequisites

articles/ai-foundry/foundry-local/how-to/how-to-use-langchain-with-foundry-local.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,8 @@ zone_pivot_groups: foundry-local-sdk
1717

1818
# Build a translation application with LangChain
1919

20+
[!INCLUDE [foundry-local-preview](./../includes/foundry-local-preview.md)]
21+
2022
This tutorial shows you how to create an application using the Foundry Local SDK and [LangChain](https://www.langchain.com/langchain). In this tutorial, you build a translation application that translates text from one language to another that uses a local model.
2123

2224
::: zone pivot="programming-language-python"
Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,14 @@
1+
---
2+
title: include file
3+
description: include file
4+
author: jonburchel
5+
ms.author: jburchel
6+
ms.service: azure-ai-foundry
7+
ms.topic: include
8+
ms.date: 05/19/2025
9+
ms.custom: include file
10+
---
11+
12+
> [!IMPORTANT]
13+
> - Foundry Local is available in preview. Public preview releases provide early access to features that are in active deployment.
14+
> - Features, approaches, and processes can change or have limited capabilities, before General Availability (GA).

articles/ai-foundry/foundry-local/includes/integrate-examples/javascript.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -32,15 +32,15 @@ import { FoundryLocalManager } from "foundry-local-sdk";
3232
// to your end-user's device.
3333
// TIP: You can find a list of available models by running the
3434
// following command in your terminal: `foundry model list`.
35-
const modelAlias = "deepseek-r1-1.5b";
35+
const alias = "deepseek-r1-1.5b";
3636

3737
// Create a FoundryLocalManager instance. This will start the Foundry
3838
// Local service if it is not already running.
3939
const foundryLocalManager = new FoundryLocalManager()
4040

4141
// Initialize the manager with a model. This will download the model
4242
// if it is not already present on the user's device.
43-
const modelInfo = await foundryLocalManager.init(modelAlias)
43+
const modelInfo = await foundryLocalManager.init(alias)
4444
console.log("Model Info:", modelInfo)
4545

4646
const openai = new OpenAI({
@@ -83,15 +83,15 @@ import { FoundryLocalManager } from "foundry-local-sdk";
8383
// to your end-user's device.
8484
// TIP: You can find a list of available models by running the
8585
// following command in your terminal: `foundry model list`.
86-
const modelAlias = "deepseek-r1-1.5b";
86+
const alias = "deepseek-r1-1.5b";
8787

8888
// Create a FoundryLocalManager instance. This will start the Foundry
8989
// Local service if it is not already running.
9090
const foundryLocalManager = new FoundryLocalManager()
9191

9292
// Initialize the manager with a model. This will download the model
9393
// if it is not already present on the user's device.
94-
const modelInfo = await foundryLocalManager.init(modelAlias)
94+
const modelInfo = await foundryLocalManager.init(alias)
9595
console.log("Model Info:", modelInfo)
9696

9797
const openai = new OpenAI({
@@ -133,15 +133,15 @@ import { FoundryLocalManager } from "foundry-local-sdk";
133133
// to your end-user's device.
134134
// TIP: You can find a list of available models by running the
135135
// following command in your terminal: `foundry model list`.
136-
const modelAlias = "deepseek-r1-1.5b";
136+
const alias = "deepseek-r1-1.5b";
137137

138138
// Create a FoundryLocalManager instance. This will start the Foundry
139139
// Local service if it is not already running.
140140
const foundryLocalManager = new FoundryLocalManager()
141141

142142
// Initialize the manager with a model. This will download the model
143143
// if it is not already present on the user's device.
144-
const modelInfo = await foundryLocalManager.init(modelAlias)
144+
const modelInfo = await foundryLocalManager.init(alias)
145145
console.log("Model Info:", modelInfo)
146146

147147
async function queryModel() {
@@ -176,15 +176,15 @@ import { FoundryLocalManager } from "foundry-local-sdk";
176176
// to your end-user's device.
177177
// TIP: You can find a list of available models by running the
178178
// following command in your terminal: `foundry model list`.
179-
const modelAlias = "deepseek-r1-1.5b";
179+
const alias = "deepseek-r1-1.5b";
180180

181181
// Create a FoundryLocalManager instance. This will start the Foundry
182182
// Local service if it is not already running.
183183
const foundryLocalManager = new FoundryLocalManager()
184184

185185
// Initialize the manager with a model. This will download the model
186186
// if it is not already present on the user's device.
187-
const modelInfo = await foundryLocalManager.init(modelAlias)
187+
const modelInfo = await foundryLocalManager.init(alias)
188188
console.log("Model Info:", modelInfo)
189189

190190
async function streamWithFetch() {

articles/ai-foundry/foundry-local/includes/sdk-reference/javascript.md

Lines changed: 27 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -33,12 +33,21 @@ Available options:
3333
- `serviceUrl`: Base URL of the Foundry Local service
3434
- `fetch`: (optional) Custom fetch implementation for environments like Node.js
3535

36+
### A note on aliases
37+
38+
Many methods outlined in this reference have an `aliasOrModelId` parameter in the signature. You can pass into the method either an **alias** or **model ID** as a value. Using an alias will:
39+
40+
- Select the *best model* for the available hardware. For example, if a Nvidia CUDA GPU is available, Foundry Local selects the CUDA model. If a supported NPU is available, Foundry Local selects the NPU model.
41+
- Allow you to use a shorter name without needing to remember the model ID.
42+
43+
> [!TIP]
44+
> We recommend passing into the `aliasOrModelId` parameter an **alias** because when you deploy your application, Foundry Local acquires the best model for the end user's machine at run-time.
3645
3746
### Service Management
3847

3948
| Method | Signature | Description |
4049
|-----------------------|---------------------------|--------------------------------------------------|
41-
| `init()` | `(modelAliasOrId?: string) => Promise<void>` | Initializes the SDK and optionally loads a model. |
50+
| `init()` | `(aliasOrModelId?: string) => Promise<void>` | Initializes the SDK and optionally loads a model. |
4251
| `isServiceRunning()` | `() => Promise<boolean>` | Checks if the Foundry Local service is running. |
4352
| `startService()` | `() => Promise<void>` | Starts the Foundry Local service. |
4453
| `serviceUrl` | `string` | The base URL of the Foundry Local service. |
@@ -52,24 +61,24 @@ Available options:
5261
|---------------------------|---------------------------------------------------------------------------|--------------------------------------------------|
5362
| `listCatalogModels()` | `() => Promise<FoundryModelInfo[]>` | Lists all available models in the catalog. |
5463
| `refreshCatalog()` | `() => Promise<void>` | Refreshes the model catalog. |
55-
| `getModelInfo()` | `(modelAliasOrId: string, throwOnNotFound = false) => Promise<FoundryModelInfo \| null>` | Gets model info by alias or ID. |
64+
| `getModelInfo()` | `(aliasOrModelId: string, throwOnNotFound = false) => Promise<FoundryModelInfo \| null>` | Gets model info by alias or ID. |
5665

5766

5867
### Cache Management
5968

6069
| Method | Signature | Description |
6170
|---------------------------|---------------------------------------------------|--------------------------------------------------|
6271
| `getCacheLocation()` | `() => Promise<string>` | Returns the model cache directory path. |
63-
| `listLocalModels()` | `() => Promise<FoundryModelInfo[]>` | Lists models downloaded to the local cache. |
72+
| `listCachedModels()` | `() => Promise<FoundryModelInfo[]>` | Lists models downloaded to the local cache. |
6473

6574

6675
### Model Management
6776

6877
| Method | Signature | Description |
6978
|-------------------------------|---------------------------------------------------------------------------|--------------------------------------------------|
70-
| `downloadModel()` | `(modelAliasOrId: string, force = false, onProgress?) => Promise<FoundryModelInfo>` | Downloads a model to the local cache. |
71-
| `loadModel()` | `(modelAliasOrId: string, ttl = 600) => Promise<FoundryModelInfo>` | Loads a model into the inference server. |
72-
| `unloadModel()` | `(modelAliasOrId: string, force = false) => Promise<void>` | Unloads a model from the inference server. |
79+
| `downloadModel()` | `(aliasOrModelId: string, token?: string, force = false, onProgress?) => Promise<FoundryModelInfo>` | Downloads a model to the local cache. |
80+
| `loadModel()` | `(aliasOrModelId: string, ttl = 600) => Promise<FoundryModelInfo>` | Loads a model into the inference server. |
81+
| `unloadModel()` | `(aliasOrModelId: string, force = false) => Promise<void>` | Unloads a model from the inference server. |
7382
| `listLoadedModels()` | `() => Promise<FoundryModelInfo[]>` | Lists all models currently loaded in the service.|
7483

7584
## Example Usage
@@ -83,12 +92,12 @@ import { FoundryLocalManager } from "foundry-local-sdk";
8392
// to your end-user's device.
8493
// TIP: You can find a list of available models by running the
8594
// following command in your terminal: `foundry model list`.
86-
const modelAlias = "deepseek-r1-1.5b";
95+
const alias = "deepseek-r1-1.5b";
8796

8897
const manager = new FoundryLocalManager()
8998

9099
// Initialize the SDK and optionally load a model
91-
const modelInfo = await manager.init(modelAlias)
100+
const modelInfo = await manager.init(alias)
92101
console.log("Model Info:", modelInfo)
93102

94103
// Check if the service is running
@@ -99,17 +108,17 @@ console.log(`Service running: ${isRunning}`)
99108
const catalog = await manager.listCatalogModels()
100109

101110
// Download and load a model
102-
await manager.downloadModel(modelAlias)
103-
await manager.loadModel(modelAlias)
111+
await manager.downloadModel(alias)
112+
await manager.loadModel(alias)
104113

105114
// List models in cache
106-
const localModels = await manager.listLocalModels()
115+
const localModels = await manager.listCachedModels()
107116

108117
// List loaded models
109118
const loaded = await manager.listLoadedModels()
110119

111120
// Unload a model
112-
await manager.unloadModel(modelAlias)
121+
await manager.unloadModel(alias)
113122
```
114123

115124
---
@@ -132,15 +141,15 @@ import { FoundryLocalManager } from "foundry-local-sdk";
132141
// to your end-user's device.
133142
// TIP: You can find a list of available models by running the
134143
// following command in your terminal: `foundry model list`.
135-
const modelAlias = "deepseek-r1-1.5b";
144+
const alias = "deepseek-r1-1.5b";
136145

137146
// Create a FoundryLocalManager instance. This will start the Foundry
138147
// Local service if it is not already running.
139148
const foundryLocalManager = new FoundryLocalManager()
140149

141150
// Initialize the manager with a model. This will download the model
142151
// if it is not already present on the user's device.
143-
const modelInfo = await foundryLocalManager.init(modelAlias)
152+
const modelInfo = await foundryLocalManager.init(alias)
144153
console.log("Model Info:", modelInfo)
145154

146155
const openai = new OpenAI({
@@ -199,15 +208,15 @@ const endpoint = "ENDPOINT"
199208

200209
const manager = new FoundryLocalManager({serviceUrl: endpoint})
201210

202-
const modelAlias = 'deepseek-r1-1.5b'
211+
const alias = 'deepseek-r1-1.5b'
203212

204213
// Get all available models
205214
const catalog = await manager.listCatalogModels()
206215
console.log("Available models in catalog:", catalog)
207216

208217
// Download and load a specific model
209-
await manager.downloadModel(modelAlias)
210-
await manager.loadModel(modelAlias)
218+
await manager.downloadModel(alias)
219+
await manager.loadModel(alias)
211220

212221
// View models in your local cache
213222
const localModels = await manager.listLocalModels()
@@ -218,5 +227,5 @@ const loaded = await manager.listLoadedModels()
218227
console.log("Loaded models in inference service:", loaded)
219228

220229
// Unload a model when finished
221-
await manager.unloadModel(modelAlias)
230+
await manager.unloadModel(alias)
222231
```

0 commit comments

Comments
 (0)