You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Foundry Local is an on-device AI inference solution offering performance, privacy, customization, and cost advantages. It integrates seamlessly into your existing workflows and applications through an intuitive CLI, SDK, and REST API.
23
+
Foundry Local is an on-device AI inference solution that provides performance, privacy, customization, and cost benefits. It integrates with your workflows and applications through a CLI, SDK, and REST API.
23
24
24
25
## Key features
25
26
26
-
-**On-Device Inference**: Run models locally on your own hardware, reducing your costs while keeping all your data on your device.
27
+
-**On-Device inference**: Run models locally to reduce costs and keep data on your device.
27
28
28
-
-**Model Customization**: Select from preset models or use your own to meet specific requirements and use cases.
29
+
-**Model customization**: Select a preset model or use your own to meet specific needs.
29
30
30
-
-**Cost Efficiency**: Eliminate recurring cloud service costs by using your existing hardware, making AI more accessible.
31
+
-**Cost efficiency**: Use existing hardware to eliminate recurring cloud costs and make AI more accessible.
31
32
32
-
-**Seamless Integration**: Connect with your applications through an SDK, API endpoints, or the CLI, with easy scaling to Azure AI Foundry as your needs grow.
33
+
-**Seamless integration**: Integrate with your apps through the SDK, API endpoints, or CLI, and scale to Azure AI Foundry as your needs grow.
33
34
34
35
## Use cases
35
36
36
-
Foundry Local is ideal for scenarios where:
37
+
Foundry Local is ideal when you need to:
37
38
38
-
-You want to keep sensitive data on your device.
39
-
-You need to operate in environments with limited or no internet connectivity.
40
-
-You want to reduce cloud inference costs.
41
-
-You need low-latency AI responses for real-time applications.
42
-
-You want to experiment with AI models before deploying to a cloud environment.
39
+
-Keep sensitive data on your device
40
+
-Operate in limited or offline environments
41
+
-Reduce cloud inference costs
42
+
-Get lowlatency AI responses for real-time applications
43
+
-Experiment with AI models before you deploy to the cloud
43
44
44
-
## Do I need an Azure subscription?
45
+
## Do I need an Azure subscription
45
46
46
-
No - Foundry Local doesn't require an Azure subscription. It runs on your local hardware, allowing you to use your existing infrastructure without the need for cloud services.
47
+
No. Foundry Local runs on your hardware, letting you use your existing infrastructure without cloud services.
47
48
48
49
## Frequently asked questions
49
50
50
-
### Do I need special drivers for NPU acceleration?
51
+
### Do I need special drivers for NPU acceleration
51
52
52
-
Yes, depending on your NPU hardware, you need to install the appropriate drivers:
53
+
Install the driver for your NPU hardware:
53
54
54
-
-**Intel NPU**: Install the [Intel NPU driver](https://www.intel.com/content/www/us/en/download/794734/intel-npu-driver-windows.html) to enable NPU acceleration on Windows.
55
+
- Intel NPU: Install the [Intel NPU driver](https://www.intel.com/content/www/us/en/download/794734/intel-npu-driver-windows.html) to enable NPU acceleration on Windows.
55
56
56
-
-**Qualcomm NPU**: Install the [Qualcomm NPU driver](https://softwarecenter.qualcomm.com/catalog/item/QHND) to enable NPU acceleration. If you encounter the error `Qnn error code 5005: "Failed to load from EpContext model. qnn_backend_manager."`, this typically indicates an outdated driver or NPU resource conflicts. Try rebooting to clear NPU resource conflicts, especially after using Windows Copilot+ features.
57
+
- Qualcomm NPU: Install the [Qualcomm NPU driver](https://softwarecenter.qualcomm.com/catalog/item/QHND) to enable NPU acceleration. If you see the error `Qnn error code 5005: Failed to load from EpContext model. qnn_backend_manager.`, it likely indicates an outdated driver or an NPU resource conflict. Reboot to clear the conflict, especially after using Windows Copilot+ features.
57
58
58
-
Foundry Local automatically detects and uses available NPU hardware once the proper drivers are installed.
59
+
After you install the drivers, Foundry Local automatically detects and uses the NPU.
59
60
60
-
## Getting started
61
+
## Get started
61
62
62
-
Install and run your first model by following the [Get started with Foundry Local](get-started.md) guide. This guide helps you set up Foundry Local, discover available models, and launch your first local AI model.
63
+
Follow the [Get started with Foundry Local](get-started.md) guide to set up Foundry Local, discover models, and run your first local AI model.
63
64
64
-
## Next steps
65
+
## Related content
65
66
66
67
-[Get started with Foundry Local](get-started.md)
67
-
-[How to compile Hugging Face models to run on Foundry Local](how-to/how-to-compile-hugging-face-models.md)
68
+
-[Compile Hugging Face models for Foundry Local](how-to/how-to-compile-hugging-face-models.md)
0 commit comments