Skip to content

Commit b5480ae

Browse files
committed
fixed review2 feedback.
1 parent 0ddb5b7 commit b5480ae

10 files changed

+23
-34
lines changed

articles/ai-foundry-local/concepts/foundry-local-architecture.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -80,7 +80,7 @@ The model cache stores downloaded AI models locally on your device, which ensure
8080
Before models can be used with Foundry Local, they must be compiled and optimized in the [ONNX](https://onnx.ai) format. Microsoft provides a selection of published models in the Azure AI Foundry Model Catalog that are already optimized for Foundry Local. However, you aren't limited to those models - by using [Olive](https://microsoft.github.io/Olive/). Olive is a powerful framework for preparing AI models for efficient inference. It converts models into the ONNX format, optimizes their graph structure, and applies techniques like quantization to improve performance on local hardware.
8181

8282
> [!TIP]
83-
> To learn more about compiling models for Foundry Local, read [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-huggingface-models.md).
83+
> To learn more about compiling models for Foundry Local, read [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-hugging-face-models.md).
8484
8585
### Hardware abstraction layer
8686

articles/ai-foundry-local/get-started.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -93,5 +93,5 @@ foundry cache --help
9393
- [Explore the Foundry Local documentation](index.yml)
9494
- [Learn about best practices and troubleshooting](reference/reference-best-practice.md)
9595
- [Explore the Foundry Local API reference](reference/reference-catalog-api.md)
96-
- [Learn how to compile Hugging Face models](how-to/how-to-compile-huggingface-models.md)
96+
- [Learn how to compile Hugging Face models](how-to/how-to-compile-hugging-face-models.md)
9797

articles/ai-foundry-local/how-to/how-to-compile-huggingface-models.md renamed to articles/ai-foundry-local/how-to/how-to-compile-hugging-face-models.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: How to compile HuggingFace models to run on Foundry Local
2+
title: How to compile Hugging Face models to run on Foundry Local
33
titleSuffix: Foundry Local
4-
description: Learn how to compile and run HuggingFace models with Foundry Local.
4+
description: Learn how to compile and run Hugging Face models with Foundry Local.
55
manager: scottpolly
66
ms.service: azure-ai-foundry
77
ms.custom: build-2025
@@ -11,7 +11,7 @@ ms.author: samkemp
1111
author: samuel100
1212
---
1313

14-
# How to compile HuggingFace models to run on Foundry Local
14+
# How to compile Hugging Face models to run on Foundry Local
1515

1616
Foundry Local runs ONNX models on your device with high performance. While the model catalog offers _out-of-the-box_ precompiled options, you can use any model in the ONNX format.
1717

@@ -21,7 +21,7 @@ This guide shows you how to:
2121

2222
> [!div class="checklist"]
2323
>
24-
> - **Convert and optimize** models from HuggingFace to run in Foundry Local. You'll use the `Llama-3.2-1B-Instruct` model as an example, but you can use any generative AI model from HuggingFace.
24+
> - **Convert and optimize** models from Hugging Face to run in Foundry Local. You'll use the `Llama-3.2-1B-Instruct` model as an example, but you can use any generative AI model from Hugging Face.
2525
> - **Run** your optimized models with Foundry Local
2626
2727
## Prerequisites
@@ -49,9 +49,9 @@ pip install olive-ai[auto-opt]
4949
> [!TIP]
5050
> For best results, install Olive in a virtual environment using [venv](https://docs.python.org/3/library/venv.html) or [conda](https://www.anaconda.com/docs/getting-started/miniconda/main).
5151
52-
## Sign in to HuggingFace
52+
## Sign in to Hugging Face
5353

54-
You optimize the `Llama-3.2-1B-Instruct` model, which requires HuggingFace authentication:
54+
You optimize the `Llama-3.2-1B-Instruct` model, which requires Hugging Face authentication:
5555

5656
### [Bash](#tab/Bash)
5757

@@ -68,7 +68,7 @@ huggingface-cli login
6868
---
6969

7070
> [!NOTE]
71-
> You must first [create a HuggingFace token](https://huggingface.co/docs/hub/security-tokens) and [request model access](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) before proceeding.
71+
> You must first [create a Hugging Face token](https://huggingface.co/docs/hub/security-tokens) and [request model access](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) before proceeding.
7272
7373
## Compile the model
7474

@@ -113,15 +113,15 @@ The command uses the following parameters:
113113

114114
| Parameter | Description |
115115
| -------------------- | --------------------------------------------------------------------------------- |
116-
| `model_name_or_path` | Model source: HuggingFace ID, local path, or Azure AI Model registry ID |
116+
| `model_name_or_path` | Model source: Hugging Face ID, local path, or Azure AI Model registry ID |
117117
| `output_path` | Where to save the optimized model |
118118
| `device` | Target hardware: `cpu`, `gpu`, or `npu` |
119119
| `provider` | Execution provider (for example, `CPUExecutionProvider`, `CUDAExecutionProvider`) |
120120
| `precision` | Model precision: `fp16`, `fp32`, `int4`, or `int8` |
121121
| `use_ort_genai` | Creates inference configuration files |
122122

123123
> [!TIP]
124-
> If you have a local copy of the model, you can use a local path instead of the HuggingFace ID. For example, `--model_name_or_path models/llama-3.2-1B-Instruct`. Olive handles the conversion, optimization, and quantization automatically.
124+
> If you have a local copy of the model, you can use a local path instead of the Hugging Face ID. For example, `--model_name_or_path models/llama-3.2-1B-Instruct`. Olive handles the conversion, optimization, and quantization automatically.
125125
126126
### Step 2: Rename the output model
127127

@@ -159,10 +159,10 @@ Foundry Local requires a chat template JSON file called `inference_model.json` i
159159
}
160160
```
161161

162-
To create the chat template file, you can use the `apply_chat_template` method from the HuggingFace library:
162+
To create the chat template file, you can use the `apply_chat_template` method from the Hugging Face library:
163163

164164
> [!NOTE]
165-
> The following example uses the Python HuggingFace library to create a chat template. The HuggingFace library is a dependency for Olive, so if you're using the same Python virtual environment you don't need to install. If you're using a different environment, install the library with `pip install transformers`.
165+
> The following example uses the Python Hugging Face library to create a chat template. The Hugging Face library is a dependency for Olive, so if you're using the same Python virtual environment you don't need to install. If you're using a different environment, install the library with `pip install transformers`.
166166
167167
```python
168168
# generate_inference_model.py

articles/ai-foundry-local/how-to/integrate-with-inference-sdks.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,5 +48,5 @@ When Foundry Local is running, it exposes an OpenAI-compatible REST API endpoint
4848

4949
## Next steps
5050

51-
- [How to compile Hugging Face models to run on Foundry Local](how-to-compile-huggingface-models.md)
51+
- [How to compile Hugging Face models to run on Foundry Local](how-to-compile-hugging-face-models.md)
5252
- [Explore the Foundry Local CLI reference](../reference/reference-cli.md)

articles/ai-foundry-local/index.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ landingContent:
4444
- text: Integrate with Inferencing SDKs
4545
url: how-to/integrate-with-inference-sdks.md
4646
- text: Compile Hugging Face models to run on Foundry Local
47-
url: how-to/how-to-compile-huggingface-models.md
47+
url: how-to/how-to-compile-hugging-face-models.md
4848
# Card
4949
- title: Reference
5050
linkLists:

articles/ai-foundry-local/reference/reference-catalog-api.md

Lines changed: 1 addition & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
---
1+
22
title: Catalog API Reference
33
titleSuffix: Foundry Local
44
description: Complete reference guide for the Foundry Local Model Catalog API.
@@ -9,7 +9,6 @@ ms.topic: reference
99
ms.date: 05/02/2025
1010
ms.author: maanavdalal
1111
author: maanavd
12-
---
1312

1413
# Catalog API Reference
1514

@@ -23,24 +22,19 @@ https://<catalog provider URI>/<provider subpath>
2322

2423
Replace `<catalog provider URI>` and `<provider subpath>` with your specific catalog hosting information.
2524

26-
---
27-
2825
## Available APIs
2926

3027
- **Get Model Details** - Retrieve information about a specific model
3128
- **Get Publisher Details** - Access publisher information
3229
- **List Publishers** - View all available publishers
3330
- **List Models** - Browse available models
3431

35-
---
36-
3732
## Authorization
3833

3934
All endpoints must support:
4035

4136
- Anonymous access (no authentication required)
4237

43-
---
4438

4539
## Get Model Details
4640

@@ -60,7 +54,6 @@ GET https://<catalog provider URI>/<provider subpath>/models/{modelName}/version
6054
curl -X GET --location "https://<catalog provider URI>/<provider subpath>/models/Phi-3-mini-128k-instruct/version/12"
6155
```
6256

63-
---
6457

6558
## Get Publisher Details
6659

@@ -78,7 +71,6 @@ GET https://<catalog provider URI>/<provider subpath>/publishers/{publisherName}
7871
curl -X GET --location "https://<catalog provider URI>/<provider subpath>/publishers/contoso"
7972
```
8073

81-
---
8274

8375
## List Publishers
8476

@@ -98,7 +90,6 @@ curl -X POST --location "https://<catalog provider URI>/<provider subpath>/publi
9890
--data '{"continuationToken": ""}'
9991
```
10092

101-
---
10293

10394
## List Models
10495

@@ -145,7 +136,6 @@ curl -X POST --location "https://<catalog provider URI>/<provider subpath>/model
145136
}'
146137
```
147138

148-
---
149139

150140
## Filterable Fields
151141

@@ -183,6 +173,5 @@ You can order results by the following fields:
183173
- `displayName`
184174
- `publisher`
185175

186-
---
187176

188177
For more information, refer to your catalog provider's API documentation.

articles/ai-foundry-local/toc.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ items:
1616
- name: Integrate with Inferencing SDKs
1717
href: how-to/integrate-with-inference-sdks.md
1818
- name: Compile Hugging Face models to run on Foundry Local
19-
href: how-to/how-to-compile-huggingface-models.md
19+
href: how-to/how-to-compile-hugging-face-models.md
2020
- name: Tutorials
2121
expanded: true
2222
items:

articles/ai-foundry-local/tutorials/chat-application-with-open-web-ui.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,10 +42,10 @@ Before you start this tutorial, you need:
4242

4343
4. **Connect Open Web UI to Foundry Local**:
4444

45-
1. Click **Settings** in the navigation menu
45+
1. Select **Settings** in the navigation menu
4646
2. Select **Connections**
47-
3. Click **Manage Direct Connections**
48-
4. Click the **+** icon to add a connection
47+
3. Select **Manage Direct Connections**
48+
4. Select the **+** icon to add a connection
4949
5. Enter `http://localhost:5272/v1` for the URL
5050
6. Type any value (like `test`) for the API Key, since it cannot be empty
5151
7. Save your connection
@@ -60,4 +60,4 @@ That's it! You're now chatting with an AI model running entirely on your local d
6060
## Next steps
6161

6262
- [Build an application with LangChain](use-langchain-with-foundry-local.md)
63-
- [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-huggingface-models.md)
63+
- [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-hugging-face-models.md)

articles/ai-foundry-local/tutorials/use-langchain-with-foundry-local.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -82,4 +82,4 @@ print(ai_msg)
8282
## Next steps
8383

8484
- Explore the [LangChain documentation](https://python.langchain.com/docs/introduction) for more advanced features and capabilities.
85-
- [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-huggingface-models.md)
85+
- [How to compile Hugging Face models to run on Foundry Local](../how-to/how-to-compile-hugging-face-models.md)

articles/ai-foundry-local/what-is-foundry-local.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,5 +49,5 @@ Install and run your first model by following the [Get started with Foundry Loca
4949
## Next steps
5050

5151
- [Get started with Foundry Local](get-started.md)
52-
- [How to compile Hugging Face models to run on Foundry Local](how-to/how-to-compile-huggingface-models.md)
52+
- [How to compile Hugging Face models to run on Foundry Local](how-to/how-to-compile-hugging-face-models.md)
5353

0 commit comments

Comments
 (0)