Skip to content

Commit b3f8788

Browse files
committed
Squashed commit of the following:
commit 43cf3ee Merge: 495d9f5 a04ff9e Author: Diana Richards <[email protected]> Date: Mon May 12 12:57:04 2025 -0500 Merge pull request #4739 from samuel100/samuel100/foundry-local-sdk-updates Samuel100/foundry local sdk updates commit a04ff9e Author: Diana Richards <[email protected]> Date: Mon May 12 12:51:46 2025 -0500 Apply suggestions from code review commit 495d9f5 Merge: f756f03 9bbf7d1 Author: Alma Jenks <[email protected]> Date: Mon May 12 17:06:03 2025 +0000 Merge branch 'main' into release-build-foundry-local commit 6218709 Author: samuel100 <[email protected]> Date: Mon May 12 17:12:03 2025 +0100 remove azure blob requirement commit abee012 Author: samuel100 <[email protected]> Date: Mon May 12 15:42:03 2025 +0100 fix link commit 51aa70b Author: samuel100 <[email protected]> Date: Mon May 12 15:30:31 2025 +0100 updated with guidance on dynamically allocated port. commit 4e89f2c Author: samuel100 <[email protected]> Date: Mon May 12 15:21:45 2025 +0100 add example of blob path commit 8ae019b Author: samuel100 <[email protected]> Date: Mon May 12 14:54:08 2025 +0100 fixed the dynamic allocation of endpoint commit d77298a Author: samuel100 <[email protected]> Date: Mon May 12 14:43:57 2025 +0100 remove c# code for FL commit 5578a85 Author: samuel100 <[email protected]> Date: Mon May 12 14:40:17 2025 +0100 address feedback commit 81ad354 Author: samuel100 <[email protected]> Date: Sun May 11 18:16:23 2025 +0100 more details on the request and response. commit dac2d1d Author: samuel100 <[email protected]> Date: Sun May 11 17:54:31 2025 +0100 fix embeddings api commit 98ba922 Author: samuel100 <[email protected]> Date: Sun May 11 17:51:29 2025 +0100 include foundry/list api commit 031902f Author: samuel100 <[email protected]> Date: Sun May 11 17:37:06 2025 +0100 typo commit 5aedc3d Author: samuel100 <[email protected]> Date: Sun May 11 17:35:30 2025 +0100 reference catalog API updated to latest commit 8adff40 Author: samuel100 <[email protected]> Date: Sun May 11 17:13:21 2025 +0100 added a bit of commentary on the sdk examples. commit b785913 Author: samuel100 <[email protected]> Date: Sun May 11 16:51:21 2025 +0100 intro examples commit 11e9150 Author: samuel100 <[email protected]> Date: Sun May 11 16:50:08 2025 +0100 reference sdk tidy and update commit b2d6873 Author: samuel100 <[email protected]> Date: Sun May 11 15:54:16 2025 +0100 match output to model in article commit 781470a Author: samuel100 <[email protected]> Date: Sun May 11 15:47:23 2025 +0100 fix typo commit 9ee05e6 Author: samuel100 <[email protected]> Date: Sun May 11 15:39:20 2025 +0100 fix warnings commit 251fd50 Author: samuel100 <[email protected]> Date: Sun May 11 15:25:41 2025 +0100 fix additional zone-end commit 5df693c Author: samuel100 <[email protected]> Date: Sun May 11 15:22:26 2025 +0100 collapse tutorials into how-tos commit 987a4e4 Author: samuel100 <[email protected]> Date: Sun May 11 14:54:06 2025 +0100 fix zone groups commit d5b34ea Author: samuel100 <[email protected]> Date: Sun May 11 14:47:41 2025 +0100 fl sdk added to inferencing sdk docs commit b635fb7 Author: samuel100 <[email protected]> Date: Sun May 11 12:48:42 2025 +0100 javascript sdks updated with fl sdk commit 3cc46fa Author: samuel100 <[email protected]> Date: Sun May 11 11:59:55 2025 +0100 fix tip commit d23ac13 Author: samuel100 <[email protected]> Date: Sun May 11 11:59:39 2025 +0100 change winget to a tip commit 8375858 Author: samuel100 <[email protected]> Date: Sun May 11 11:54:57 2025 +0100 include expected output gif in quickstart. commit 7130bee Author: samuel100 <[email protected]> Date: Sun May 11 11:07:17 2025 +0100 updates to javascript langchain with foundry sdk commit 0c1e202 Author: samuel100 <[email protected]> Date: Sun May 11 10:03:42 2025 +0100 acrolinx to 100 commit 9eefe73 Author: samuel100 <[email protected]> Date: Sun May 11 09:59:33 2025 +0100 move files to new location commit 99dc3e4 Merge: a7fb484 f756f03 Author: samuel100 <[email protected]> Date: Sun May 11 09:58:52 2025 +0100 Merge branch 'release-build-foundry-local' of https://github.com/samuel100/azure-ai-docs-pr into samuel100/foundry-local-sdk-updates commit f756f03 Merge: 5537cc0 55f7f22 Author: Colin Cooper <[email protected]> Date: Sun May 11 09:38:09 2025 +0100 Merge pull request #4735 from jonburchel/release-build-foundry-local-updates [SCOPED] Moving Foundry Local files to /articles/ai-foundry/foundry-local folder. commit 55f7f22 Author: Jon Burchel <[email protected]> Date: Sat May 10 22:09:14 2025 -0400 Moving Foundry Local files to /articles/ai-foundry/foundry-local folder. commit 5537cc0 Merge: 38ca6f2 5d02c9d Author: Anna Huff <[email protected]> Date: Sat May 10 16:31:37 2025 -0600 Merge pull request #3591 from samuel100/samuel100/foundry-local Foundry Local commit a7fb484 Merge: 0602daa 5fb9e73 Author: samuel100 <[email protected]> Date: Sat May 10 14:25:30 2025 +0100 Merge branch 'samuel100/foundry-local-sdk-updates' of https://github.com/samuel100/azure-ai-docs-pr into samuel100/foundry-local-sdk-updates commit 0602daa Author: samuel100 <[email protected]> Date: Sat May 10 14:25:11 2025 +0100 python sdk doc on langchain added. commit 5d02c9d Author: samuel100 <[email protected]> Date: Sat May 10 13:05:26 2025 +0100 fixed the extra line commit 1f17d8f Author: samuel100 <[email protected]> Date: Sat May 10 12:55:44 2025 +0100 fixed missing --- in attributes commit 5fb9e73 Merge: 4d1f25e 38ca6f2 Author: Samuel Kemp <[email protected]> Date: Sat May 10 12:51:34 2025 +0100 Merge branch 'MicrosoftDocs:main' into samuel100/foundry-local-sdk-updates commit b5480ae Author: samuel100 <[email protected]> Date: Sat May 10 12:46:17 2025 +0100 fixed review2 feedback. commit 0ddb5b7 Merge: 4d1f25e dbb2436 Author: Maanav Dalal <[email protected]> Date: Fri May 9 18:50:19 2025 -0400 Merge pull request #13 from samuel100/fixing_merge_errors Fixed suggestions. commit dbb2436 Author: MaanavD <[email protected]> Date: Fri May 9 18:48:52 2025 -0400 Fixed suggestions. commit 4d1f25e Author: samuel100 <[email protected]> Date: Fri May 9 21:33:25 2025 +0100 fixed ms.topic commit 43e1e38 Author: samuel100 <[email protected]> Date: Fri May 9 21:31:26 2025 +0100 review feedback addressed commit 0bbab1c Author: samuel100 <[email protected]> Date: Fri May 9 15:06:25 2025 +0100 another fix commit a384a1e Author: samuel100 <[email protected]> Date: Fri May 9 14:59:58 2025 +0100 fix pivot group id commit 6d6f215 Merge: f4b1809 f648e4b Author: samuel100 <[email protected]> Date: Fri May 9 14:52:07 2025 +0100 Merge branch 'main' of https://github.com/samuel100/azure-ai-docs-pr into samuel100/foundry-local commit f4b1809 Author: samuel100 <[email protected]> Date: Fri May 9 14:43:25 2025 +0100 fix warnings and model name commit dc2bd61 Author: samuel100 <[email protected]> Date: Fri May 9 14:09:33 2025 +0100 fix pivot groups commit 814f41a Author: samuel100 <[email protected]> Date: Fri May 9 10:31:26 2025 +0100 clean up, acrolinx commit 3784d52 Merge: d8051e6 32e64eb Author: Maanav Dalal <[email protected]> Date: Thu May 8 17:27:00 2025 -0700 Merge pull request #12 from samuel100/acrolinx_fixes Acrolinx fixes commit 32e64eb Author: MaanavD <[email protected]> Date: Thu May 8 17:23:26 2025 -0700 AI Linted for Acrolinx scores. Will see updates. commit 9273f1d Author: MaanavD <[email protected]> Date: Thu May 8 16:53:55 2025 -0700 Fixed all warnings from Acrolinx. commit d8051e6 Merge: c241047 6de4bd8 Author: Maanav Dalal <[email protected]> Date: Thu May 8 16:32:01 2025 -0700 Merge pull request #11 from samuel100/adding_sdk_docs Adding sdk docs. Custom models / HF models are still WIP. commit 6de4bd8 Author: MaanavD <[email protected]> Date: Thu May 8 15:47:08 2025 -0700 modelid -> id commit 8cc83fa Author: MaanavD <[email protected]> Date: Thu May 8 15:40:44 2025 -0700 Settled on a foundry manager syntax. commit f8f1a53 Author: MaanavD <[email protected]> Date: Thu May 8 15:15:47 2025 -0700 Updated samples. commit 66f78b1 Author: MaanavD <[email protected]> Date: Thu May 8 14:43:32 2025 -0700 Made models generic as api/sdk now supports that. commit 7ae0a13 Author: MaanavD <[email protected]> Date: Thu May 8 14:33:34 2025 -0700 removed another constructor. commit fd4b213 Author: MaanavD <[email protected]> Date: Thu May 8 14:25:52 2025 -0700 removed constructor. commit 0cb7978 Author: MaanavD <[email protected]> Date: Thu May 8 14:20:47 2025 -0700 Made both python and js docs aligned logically. commit 36b6c59 Author: MaanavD <[email protected]> Date: Thu May 8 14:07:18 2025 -0700 Fixed API key commit 27dfbb6 Author: MaanavD <[email protected]> Date: Thu May 8 14:05:38 2025 -0700 Updated SDK docs to reflect live SDK. commit 7ce3b61 Author: MaanavD <[email protected]> Date: Tue May 6 16:53:30 2025 -0700 Removed custom model page, made custom model tip clearer. commit 56adf9f Author: MaanavD <[email protected]> Date: Tue May 6 12:58:25 2025 -0700 Fixed more ai foundry local -> foundry local. how to models is still WIP. commit 66140d1 Author: MaanavD <[email protected]> Date: Tue May 6 12:55:47 2025 -0700 added sdk docs commit 127cefc Author: MaanavD <[email protected]> Date: Tue May 6 12:55:21 2025 -0700 added reference to SDK in the Foundry Local docs. commit c241047 Author: Maanav Dalal <[email protected]> Date: Tue May 6 12:39:20 2025 -0700 Fixing catalog instances of ai foundry local (#10) * All AI Foundry Local -> Foundry Local. * All ai-foundry-local files -> foundry-local commit 4d68121 Merge: 06d8c3d 4967f31 Author: Nat Kershaw (MSFT) <[email protected]> Date: Mon May 5 14:19:43 2025 -0700 Merge pull request #9 from samuel100/fixing_catalog Fixing catalog commit 4967f31 Author: MaanavD <[email protected]> Date: Mon May 5 14:14:59 2025 -0700 Fixed catalog issues, made auth only anonymous. commit 06d8c3d Merge: 0efb824 4484331 Author: Maanav Dalal <[email protected]> Date: Mon May 5 13:17:44 2025 -0700 Merge pull request #8 from samuel100/custom_models Changed ToC a bit, added catalog api reference. commit 4484331 Author: Maanav Dalal <[email protected]> Date: Mon May 5 13:16:26 2025 -0700 Update articles/ai-foundry-local/reference/reference-catalog-api.md Co-authored-by: Nat Kershaw (MSFT) <[email protected]> commit 77da9bc Author: Maanav Dalal <[email protected]> Date: Mon May 5 13:16:17 2025 -0700 Update articles/ai-foundry-local/reference/reference-catalog-api.md Co-authored-by: Nat Kershaw (MSFT) <[email protected]> commit 9fe338f Author: Maanav Dalal <[email protected]> Date: Mon May 5 13:16:10 2025 -0700 Update articles/ai-foundry-local/reference/reference-catalog-api.md Co-authored-by: Nat Kershaw (MSFT) <[email protected]> commit e3a1134 Author: Maanav Dalal <[email protected]> Date: Mon May 5 13:16:00 2025 -0700 Update articles/ai-foundry-local/reference/reference-catalog-api.md Co-authored-by: Nat Kershaw (MSFT) <[email protected]> commit 66475dc Author: Maanav Dalal <[email protected]> Date: Mon May 5 13:15:52 2025 -0700 Update articles/ai-foundry-local/reference/reference-catalog-api.md Co-authored-by: Nat Kershaw (MSFT) <[email protected]> commit 2d59ded Author: MaanavD <[email protected]> Date: Fri May 2 11:45:19 2025 -0700 Changed ToC a bit, added catalog api reference. commit 0efb824 Author: Elena Neroslavskaya <[email protected]> Date: Wed Apr 23 13:04:35 2025 -0700 Restore LangChain tutorial for AI Foundry Local commit 9843707 Author: Elena Neroslavskaya <[email protected]> Date: Wed Apr 23 13:02:08 2025 -0700 Add LangChain tutorial for AI Foundry Local commit 212a9be Author: Elena Neroslavskaya <[email protected]> Date: Wed Apr 23 15:48:58 2025 -0400 Rename use-langchain-with-foundrylocal to use-langchain-with-foundrylocal.md commit d77dfb5 Author: Elena Neroslavskaya <[email protected]> Date: Wed Apr 23 15:48:29 2025 -0400 Create use-langchain-with-foundrylocal LAngchain example commit 2b0877c Author: Maanav Dalal <[email protected]> Date: Mon Apr 7 12:48:39 2025 -0700 added api docs w/ warning. (#6) commit 707e248 Author: Maanav Dalal <[email protected]> Date: Tue Apr 1 06:14:30 2025 -0700 Modified get started based on Jon's feedback, moved cli commands to reference. (#5) commit 2175f87 Author: Maanav Dalal <[email protected]> Date: Tue Apr 1 06:13:53 2025 -0700 Update toc for legibility. (#4) commit 4bc4b89 Author: samuel100 <[email protected]> Date: Tue Mar 25 14:36:09 2025 +0000 added step to create inference_model.json file commit a2ff330 Author: Maanav Dalal <[email protected]> Date: Tue Mar 25 02:37:10 2025 -0700 Hf model docs + Open web UI tutorial (minus images and /v1/models endpoint). (#3) * Updated hf model docs + changed name slightly. * Made entry more user friendly. * Added open webui example. ONLY WORKS IF /v1/models ENDPOINT IS FIXED. commit d1d1457 Author: samuel100 <[email protected]> Date: Sat Mar 22 17:10:17 2025 +0000 architecture doc added commit 1801653 Author: samuel100 <[email protected]> Date: Sat Mar 22 16:00:33 2025 +0000 get started update commit b4a22ff Author: samuel100 <[email protected]> Date: Sat Mar 22 15:46:12 2025 +0000 referenced the integrate SDK article. commit fe6de37 Author: samuel100 <[email protected]> Date: Sat Mar 22 15:28:48 2025 +0000 resolved file not found issue commit 95afea3 Author: samuel100 <[email protected]> Date: Sat Mar 22 15:09:26 2025 +0000 fixed pivot issue. Change curl to REST for consistency. commit 65221e1 Author: samuel100 <[email protected]> Date: Sat Mar 22 14:45:22 2025 +0000 fix zone pivot group issue commit eddd4bc Author: Maanav Dalal <[email protected]> Date: Sat Mar 22 07:35:18 2025 -0700 Install and inference (#2) * Installation already existent, added 5 inference examples. * added curl example. * Made curl work, moved streaming to stream. * attempted adding pivots. Will have to see if it worked in PR. commit 253ca32 Author: samuel100 <[email protected]> Date: Fri Mar 21 18:02:32 2025 +0000 included examples of using REST and OpenAI SDK commit 8f1f332 Author: samuel100 <[email protected]> Date: Fri Mar 21 17:33:17 2025 +0000 changed commands to tables commit c736a17 Author: samuel100 <[email protected]> Date: Thu Mar 20 10:33:49 2025 +0000 changed qucikstars to tutorials commit 3d249b7 Author: samuel100 <[email protected]> Date: Thu Mar 20 10:19:09 2025 +0000 fixed up acrolinx commit fe1d468 Merge: 60ca2ed 22a3bd2 Author: Samuel Kemp <[email protected]> Date: Thu Mar 20 10:11:32 2025 +0000 Merge pull request #1 from MaanavD/info_and_quickstart Info and quickstart commit 22a3bd2 Author: MaanavD <[email protected]> Date: Thu Mar 20 01:06:38 2025 -0700 Added Quickstart and what is ai foundry. commit 60ca2ed Author: samuel100 <[email protected]> Date: Wed Mar 19 21:47:25 2025 +0000 fix typo commit 9e20631 Author: samuel100 <[email protected]> Date: Wed Mar 19 18:50:38 2025 +0000 fix acrolinx issues commit f9c8f57 Author: samuel100 <[email protected]> Date: Wed Mar 19 18:34:09 2025 +0000 compile models using olive doc commit a2db4ba Author: samuel100 <[email protected]> Date: Wed Mar 19 14:11:45 2025 +0000 fix warning commit f2f2312 Author: samuel100 <[email protected]> Date: Wed Mar 19 14:05:47 2025 +0000 added reference doc placeholders. commit f6821c9 Author: samuel100 <[email protected]> Date: Tue Mar 18 14:19:16 2025 +0000 foundry local init
1 parent 746d8fd commit b3f8788

23 files changed

+2722
-1
lines changed
Lines changed: 132 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,132 @@
1+
---
2+
title: Foundry Local architecture
3+
titleSuffix: Foundry Local
4+
description: Learn about the architecture and components of Foundry Local
5+
manager: scottpolly
6+
ms.service: azure-ai-foundry
7+
ms.custom: build-2025
8+
ms.topic: concept-article
9+
ms.date: 02/12/2025
10+
ms.author: samkemp
11+
author: samuel100
12+
---
13+
14+
# Foundry Local architecture
15+
16+
Foundry Local enables efficient, secure, and scalable AI model inference directly on your devices. This article explains the core components of Foundry Local and how they work together to deliver AI capabilities.
17+
18+
Key benefits of Foundry Local include:
19+
20+
> [!div class="checklist"]
21+
>
22+
> - **Low Latency**: Run models locally to minimize processing time and deliver faster results.
23+
> - **Data Privacy**: Process sensitive data locally without sending it to the cloud, helping meet data protection requirements.
24+
> - **Flexibility**: Support for diverse hardware configurations lets you choose the optimal setup for your needs.
25+
> - **Scalability**: Deploy across various devices, from laptops to servers, to suit different use cases.
26+
> - **Cost-Effectiveness**: Reduce cloud computing costs, especially for high-volume applications.
27+
> - **Offline Operation**: Work without an internet connection in remote or disconnected environments.
28+
> - **Seamless Integration**: Easily incorporate into existing development workflows for smooth adoption.
29+
30+
## Key components
31+
32+
The Foundry Local architecture consists of these main components:
33+
34+
:::image type="content" source="../media/architecture/foundry-local-arch.png" alt-text="Diagram of Foundry Local Architecture.":::
35+
36+
### Foundry Local service
37+
38+
The Foundry Local Service includes an OpenAI-compatible REST server that provides a standard interface for working with the inference engine. It's also possible to manage models over REST. Developers use this API to send requests, run models, and get results programmatically.
39+
40+
- **Endpoint**: The endpoint is *dynamically allocated* when the service starts. You can find the endpoint by running the `foundry service status` command. When using Foundry Local in your applications, we recommend using the SDK that automatically handles the endpoint for you. For more details on how to use the Foundry Local SDK, read the [Integrated inferencing SDKs with Foundry Local](../how-to/how-to-integrate-with-inference-sdks.md) article.
41+
- **Use Cases**:
42+
- Connect Foundry Local to your custom applications
43+
- Execute models through HTTP requests
44+
45+
### ONNX runtime
46+
47+
The ONNX Runtime is a core component that executes AI models. It runs optimized ONNX models efficiently on local hardware like CPUs, GPUs, or NPUs.
48+
49+
**Features**:
50+
51+
- Works with multiple hardware providers (NVIDIA, AMD, Intel, Qualcomm) and device types (NPUs, CPUs, GPUs)
52+
- Offers a consistent interface for running across models different hardware
53+
- Delivers best-in-class performance
54+
- Supports quantized models for faster inference
55+
56+
### Model management
57+
58+
Foundry Local provides robust tools for managing AI models, ensuring that they're readily available for inference and easy to maintain. Model management is handled through the **Model Cache** and the **Command-Line Interface (CLI)**.
59+
60+
#### Model cache
61+
62+
The model cache stores downloaded AI models locally on your device, which ensures models are ready for inference without needing to download them repeatedly. You can manage the cache using either the Foundry CLI or REST API.
63+
64+
- **Purpose**: Speeds up inference by keeping models locally available
65+
- **Key Commands**:
66+
- `foundry cache list`: Shows all models in your local cache
67+
- `foundry cache remove <model-name>`: Removes a specific model from the cache
68+
- `foundry cache cd <path>`: Changes the storage location for cached models
69+
70+
#### Model lifecycle
71+
72+
1. **Download**: Download models from the Azure AI Foundry model catalog and save them to your local disk.
73+
2. **Load**: Load models into the Foundry Local service memory for inference. Set a TTL (time-to-live) to control how long the model stays in memory (default: 10 minutes).
74+
3. **Run**: Execute model inference for your requests.
75+
4. **Unload**: Remove models from memory to free up resources when no longer needed.
76+
5. **Delete**: Remove models from your local cache to reclaim disk space.
77+
78+
#### Model compilation using Olive
79+
80+
Before models can be used with Foundry Local, they must be compiled and optimized in the [ONNX](https://onnx.ai) format. Microsoft provides a selection of published models in the Azure AI Foundry Model Catalog that are already optimized for Foundry Local. However, you aren't limited to those models - by using [Olive](https://microsoft.github.io/Olive/). Olive is a powerful framework for preparing AI models for efficient inference. It converts models into the ONNX format, optimizes their graph structure, and applies techniques like quantization to improve performance on local hardware.
81+
82+
> [!TIP]
83+
> To learn more about compiling models for Foundry Local, read [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-hugging-face-models.md).
84+
85+
### Hardware abstraction layer
86+
87+
The hardware abstraction layer ensures that Foundry Local can run on various devices by abstracting the underlying hardware. To optimize performance based on the available hardware, Foundry Local supports:
88+
89+
- **multiple _execution providers_**, such as NVIDIA CUDA, AMD, Qualcomm, Intel.
90+
- **multiple _device types_**, such as CPU, GPU, NPU.
91+
92+
### Developer experiences
93+
94+
The Foundry Local architecture is designed to provide a seamless developer experience, enabling easy integration and interaction with AI models.
95+
Developers can choose from various interfaces to interact with the system, including:
96+
97+
#### Command-Line Interface (CLI)
98+
99+
The Foundry CLI is a powerful tool for managing models, the inference engine, and the local cache.
100+
101+
**Examples**:
102+
103+
- `foundry model list`: Lists all available models in the local cache.
104+
- `foundry model run <model-name>`: Runs a model.
105+
- `foundry service status`: Checks the status of the service.
106+
107+
> [!TIP]
108+
> To learn more about the CLI commands, read [Foundry Local CLI Reference](../reference/reference-cli.md).
109+
110+
#### Inferencing SDK integration
111+
112+
Foundry Local supports integration with various SDKs, such as the OpenAI SDK, enabling developers to use familiar programming interfaces to interact with the local inference engine.
113+
114+
- **Supported SDKs**: Python, JavaScript, C#, and more.
115+
116+
> [!TIP]
117+
> To learn more about integrating with inferencing SDKs, read [Integrate inferencing SDKs with Foundry Local](../how-to/how-to-integrate-with-inference-sdks.md).
118+
119+
#### AI Toolkit for Visual Studio Code
120+
121+
The AI Toolkit for Visual Studio Code provides a user-friendly interface for developers to interact with Foundry Local. It allows users to run models, manage the local cache, and visualize results directly within the IDE.
122+
123+
- **Features**:
124+
- Model management: Download, load, and run models from within the IDE.
125+
- Interactive console: Send requests and view responses in real-time.
126+
- Visualization tools: Graphical representation of model performance and results.
127+
128+
## Next Steps
129+
130+
- [Get started with Foundry Local](../get-started.md)
131+
- [Integrate inferencing SDKs with Foundry Local](../how-to/how-to-integrate-with-inference-sdks.md)
132+
- [Foundry Local CLI Reference](../reference/reference-cli.md)
Lines changed: 104 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,104 @@
1+
---
2+
title: Get started with Foundry Local
3+
titleSuffix: Foundry Local
4+
description: Learn how to install, configure, and run your first AI model with Foundry Local
5+
manager: scottpolly
6+
keywords: Azure AI services, cognitive, AI models, local inference
7+
ms.service: azure-ai-foundry
8+
ms.topic: quickstart
9+
ms.date: 02/20/2025
10+
ms.reviewer: samkemp
11+
ms.author: samkemp
12+
author: samuel100
13+
ms.custom: build-2025
14+
#customer intent: As a developer, I want to get started with Foundry Local so that I can run AI models locally.
15+
---
16+
17+
# Get started with Foundry Local
18+
19+
This guide walks you through setting up Foundry Local to run AI models on your device.
20+
21+
## Prerequisites
22+
23+
Your system must meet the following requirements to run Foundry Local:
24+
25+
- **Operating System**: Windows 10 (x64), Windows 11 (x64/ARM), macOS.
26+
- **Hardware**: Minimum 8GB RAM, 3GB free disk space. Recommended 16GB RAM, 15GB free disk space.
27+
- **Network**: Internet connection for initial model download (optional for offline use)
28+
- **Acceleration (optional)**: NVIDIA GPU (2,000 series or newer), AMD GPU (6,000 series or newer), Qualcomm Snapdragon X Elite (8GB or more of memory), or Apple silicon.
29+
30+
Also, ensure you have administrative privileges to install software on your device.
31+
32+
## Quickstart
33+
34+
Get started with Foundry Local quickly:
35+
36+
1. [**Download Foundry Local Installer**](https://aka.ms/foundry-local-installer) and **install** by following the on-screen prompts.
37+
> [!TIP]
38+
> If you're installing on Windows, you can also use `winget` to install Foundry Local. Open a terminal window and run the following command:
39+
>
40+
> ```powershell
41+
> winget install Microsoft.FoundryLocal
42+
> ```
43+
1. **Run your first model** Open a terminal window and run the following command to run a model:
44+
45+
```bash
46+
foundry model run deepseek-r1-1.5b
47+
```
48+
49+
The model downloads - which can take a few minutes, depending on your internet speed - and the model runs. Once the model is running, you can interact with it using the command line interface (CLI). For example, you can ask:
50+
51+
```text
52+
Why is the sky blue?
53+
```
54+
55+
You should see a response from the model in the terminal:
56+
:::image type="content" source="media/get-started-output.png" alt-text="Screenshot of output from foundry local run command." lightbox="media/get-started-output.png":::
57+
58+
59+
> [!TIP]
60+
> You can replace `deepseek-r1-1.5b` with any model name from the catalog (see `foundry model list` for available models). Foundry Local downloads the model variant that best matches your system's hardware and software configuration. For example, if you have an NVIDIA GPU, it downloads the CUDA version of the model. If you have a Qualcomm NPU, it downloads the NPU variant. If you have no GPU or NPU, it downloads the CPU version.
61+
62+
## Explore commands
63+
64+
The Foundry CLI organizes commands into these main categories:
65+
66+
- **Model**: Commands for managing and running models.
67+
- **Service**: Commands for managing the Foundry Local service.
68+
- **Cache**: Commands for managing the local model cache (downloaded models on local disk).
69+
70+
View all available commands with:
71+
72+
```bash
73+
foundry --help
74+
```
75+
76+
To view available **model** commands, run:
77+
78+
```bash
79+
foundry model --help
80+
```
81+
To view available **service** commands, run:
82+
83+
```bash
84+
foundry service --help
85+
```
86+
87+
To view available **cache** commands, run:
88+
89+
```bash
90+
foundry cache --help
91+
```
92+
93+
> [!TIP]
94+
> For a complete guide to all CLI commands and their usage, see the [Foundry Local CLI Reference](reference/reference-cli.md).
95+
96+
97+
## Next steps
98+
99+
- [Integrate inferencing SDKs with Foundry Local](how-to/how-to-integrate-with-inference-sdks.md)
100+
- [Explore the Foundry Local documentation](index.yml)
101+
- [Learn about best practices and troubleshooting](reference/reference-best-practice.md)
102+
- [Explore the Foundry Local API reference](reference/reference-catalog-api.md)
103+
- [Learn Compile Hugging Face models](how-to/how-to-compile-hugging-face-models.md)
104+
Lines changed: 59 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,59 @@
1+
---
2+
title: Integrate Open Web UI with Foundry Local
3+
titleSuffix: Foundry Local
4+
description: Learn how to create a chat application using Foundry Local and Open Web UI
5+
manager: scottpolly
6+
keywords: Azure AI services, cognitive, AI models, local inference
7+
ms.service: azure-ai-foundry
8+
ms.topic: how-to
9+
ms.date: 02/20/2025
10+
ms.reviewer: samkemp
11+
ms.author: samkemp
12+
author: samuel100
13+
ms.custom: build-2025
14+
#customer intent: As a developer, I want to get started with Foundry Local so that I can run AI models locally.
15+
---
16+
17+
# Integrate Open Web UI with Foundry Local
18+
19+
This tutorial shows you how to create a chat application using Foundry Local and Open Web UI. When you finish, you have a working chat interface running entirely on your local device.
20+
21+
## Prerequisites
22+
23+
Before you start this tutorial, you need:
24+
25+
- **Foundry Local** installed on your computer. Read the [Get started with Foundry Local](../get-started.md) guide for installation instructions.
26+
27+
## Set up Open Web UI for chat
28+
29+
1. **Install Open Web UI** by following the instructions from the [Open Web UI GitHub repository](https://github.com/open-webui/open-webui).
30+
31+
2. **Launch Open Web UI** with this command in your terminal:
32+
33+
```bash
34+
open-webui serve
35+
```
36+
37+
3. Open your web browser and go to [http://localhost:8080](http://localhost:8080).
38+
39+
4. **Connect Open Web UI to Foundry Local**:
40+
41+
1. Select **Settings** in the navigation menu
42+
2. Select **Connections**
43+
3. Select **Manage Direct Connections**
44+
4. Select the **+** icon to add a connection
45+
5. For the **URL**, enter `http://localhost:PORT/v1` where `PORT` is replaced with the port of the Foundry Local endpoint, which you can find using the CLI command `foundry service status`. Note, that Foundry Local dynamically assigns a port, so it's not always the same.
46+
6. Type any value (like `test`) for the API Key, since it can't be empty.
47+
7. Save your connection
48+
49+
5. **Start chatting with your model**:
50+
1. Your loaded models appear in the dropdown at the top
51+
2. Select any model from the list
52+
3. Type your message in the input box at the bottom
53+
54+
That's it! You're now chatting with an AI model running entirely on your local device.
55+
56+
## Next steps
57+
58+
- [Integrate inferencing SDKs with Foundry Local](how-to-integrate-with-inference-sdks.md)
59+
- [Compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-hugging-face-models.md)

0 commit comments

Comments
 (0)