You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Azure AI Custom Vision is an image recognition service that lets you build, deploy, and improve your own **image identifier** models. An image identifier applies labels to images, according to their visual characteristics. Each label represents a classification or object. Custom Vision allows you to specify your own labels and train custom models to detect them.
20
20
21
-
> [!TIP]
22
-
> The Azure AI vision Image Analysis API, based on Florence foundational model, now supports custom models with few-shot learning capability. [Use Image Analysis 4.0](../computer-vision/how-to/model-customization.md) to create custom image identifier models using the latest technology from Azure. To migrate a Custom Vision project to the new Image Analysis 4.0 system, see the [Migration guide](../computer-vision/how-to/migrate-from-custom-vision.md). To compare the two services, see the [Comparison page](./concepts/compare-alternatives.md).
21
+
23
22
24
23
25
24
You can use Custom Vision through a client library SDK, REST API, or through the [Custom Vision web portal](https://customvision.ai/). Follow a quickstart to get started.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/includes/fine-tuning-studio.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -65,9 +65,9 @@ Different model types require a different format of training data.
65
65
66
66
# [chat completion models](#tab/turbo)
67
67
68
-
The training and validation data you use **must** be formatted as a JSON Lines (JSONL) document. For `gpt-35-turbo-0613` the fine-tuning dataset must be formatted in the conversational format that is used by the [Chat completions](../how-to/chatgpt.md) API.
68
+
The training and validation data you use **must** be formatted as a JSON Lines (JSONL) document. For `gpt-35-turbo` (all versions), `gpt-4`, `gpt-4o`, and `gpt-4o-mini`, the fine-tuning dataset must be formatted in the conversational format that is used by the [Chat completions](../how-to/chatgpt.md) API.
69
69
70
-
If you would like a step-by-step walk-through of fine-tuning a `gpt-35-turbo-0613` model please refer to the [Azure OpenAI fine-tuning tutorial.](../tutorials/fine-tune.md)
70
+
If you would like a step-by-step walk-through of fine-tuning a `gpt-4o-mini` (2024-07-18) model please refer to the [Azure OpenAI fine-tuning tutorial.](../tutorials/fine-tune.md)
@@ -850,7 +850,7 @@ Alternatively, you can deploy your fine-tuned model using any of the other commo
850
850
| resource_group | The resource group name for your Azure OpenAI resource |
851
851
| resource_name | The Azure OpenAI resource name |
852
852
| model_deployment_name | The custom name for your new fine-tuned model deployment. This is the name that will be referenced in your code when making chat completion calls. |
853
-
| fine_tuned_model | Retrieve this value from your fine-tuning job results in the previous step. It will look like `gpt-4o-mini-2024-07-18.ft-b044a9d3cf9c4228b5d393567f693b83`. You'll need to add that value to the deploy_data json. |
853
+
| fine_tuned_model | Retrieve this value from your fine-tuning job results in the previous step. It will look like `gpt-4o-mini-2024-07-18.ft-0e208cf33a6a466994aff31a08aba678`. You'll need to add that value to the deploy_data json. |
"name": "<YOUR_FINE_TUNED_MODEL>", #retrieve this value from the previous call, it will look like gpt-4o-mini-2024-07-18.ft-b044a9d3cf9c4228b5d393567f693b83
877
+
"name": "<YOUR_FINE_TUNED_MODEL>", #retrieve this value from the previous call, it will look like gpt-4o-mini-2024-07-18.ft-0e208cf33a6a466994aff31a08aba678
878
878
"version": "1"
879
879
}
880
880
}
@@ -894,7 +894,7 @@ print(r.json())
894
894
895
895
You can check on your deployment progress in the Azure OpenAI Studio:
896
896
897
-
:::image type="content" source="../media/tutorials/fine-tuning/status.png" alt-text="Screenshot of the initial DataFrame table results from the CSV file." lightbox="../media/tutorials/fine-tuning/status.png":::
897
+
:::image type="content" source="../media/tutorials/fine-tuning/studio-deployment-status.png" alt-text="Screenshot of Deployment progress on Azure OpenAI Studio." lightbox="../media/tutorials/fine-tuning/studio-deployment-status.png":::
898
898
899
899
It isn't uncommon for this process to take some time to complete when dealing with deploying fine-tuned models.
900
900
@@ -968,7 +968,7 @@ You can delete the deployment in [Azure OpenAI Studio](https://oai.azure.com/),
968
968
969
969
## Troubleshooting
970
970
971
-
### How do I enable fine-tuning? Create a custom model is greyed out in Azure OpenAI Studio?
971
+
### How do I enable fine-tuning? Create a custom model is grayed out in Azure OpenAI Studio
972
972
973
973
In order to successfully access fine-tuning you need **Cognitive Services OpenAI Contributor assigned**. Even someone with high-level Service Administrator permissions would still need this account explicitly set in order to access fine-tuning. For more information please review the [role-based access control guidance](/azure/ai-services/openai/how-to/role-based-access-control#cognitive-services-openai-contributor).
Copy file name to clipboardExpand all lines: articles/search/search-get-started-rag.md
+22Lines changed: 22 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -272,6 +272,28 @@ This section uses Visual Studio Code and Python to call the chat completion APIs
272
272
273
273
You might also try the query without semantic ranking by setting `use_semantic_reranker=False`in the query parameters step. Semantic ranking can noticably improve the relevance of query results and the ability of the LLM to return useful information. Experimentation can help you decide whether it makes a difference for your content.
274
274
275
+
## Troubleshooting errors
276
+
277
+
To debug authentication errors, insert the following code before the step that calls the search engine and the LLM.
278
+
279
+
```python
280
+
import sys
281
+
import logging # Set the logging level for all azure-storage-* libraries
Rerun the query script. You should now get INFOandDEBUG statements in the output that provide more detail about the issue.
292
+
293
+
If you see output messages related to ManagedIdentityCredential and token acquisition failures, it could be that you have multiple tenants, and your Azure sign-inis using a tenant that doesn't have your search service. To get your tenant ID, search the Azure portal for "tenant properties".
294
+
295
+
Run `az login --tenant <YOUR-TENANT-ID>` at a command prompt, and then rerun the script.
296
+
275
297
## Clean up
276
298
277
299
When you're working in your own subscription, it's a good idea at the end of a project to identify whether you still need the resources you created. Resources left running can cost you money. You can delete resources individually or delete the resource group to delete the entire set of resources.
0 commit comments