Skip to content

Commit bf8aba6

Browse files
committed
2 parents f22c8ea + 647a0c9 commit bf8aba6

File tree

348 files changed

+3241
-1737
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

348 files changed

+3241
-1737
lines changed

.openpublishing.redirection.json

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4195,6 +4195,12 @@
41954195
},
41964196
{
41974197

4198+
"source_path_from_root": "/articles/cosmos-db/how-to-move-regions.md",
4199+
"redirect_url": "/azure/operational-excellence/relocation-cosmos-db",
4200+
"redirect_document_id": false
4201+
4202+
},
4203+
{
41984204
"source_path_from_root": "/articles/site-recovery/move-vaults-across-regions.md",
41994205
"redirect_url": "/azure/operational-excellence/relocation-site-recovery",
42004206
"redirect_document_id": false

articles/active-directory-b2c/faq.yml

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -274,6 +274,12 @@ sections:
274274
275275
* API connectors
276276
* Conditional Access
277+
278+
- question: |
279+
I am using rolling refresh tokens for my application and I am getting an invalid_grant error on redeeming newly acquired refresh tokens well within their set validity period. Why does this happen?
280+
answer: |
281+
While determining validity for rolling refresh tokens, B2C will consider the initial login time of the user in the application also to calculate the token validity skew. If the user haven't logged out of the application for a very long time, this skew value will exceed the validity period of the token and hence for security reasons the tokens will be considered as invalid. Hence the error. Inform the user to perform a proper logout and login back into the application and this should reset the skew. This scenario is not applicable if refresh token rolling is set as infinite rolling.
282+
277283
278284
- question: |
279285
I've revoked the refresh token using Microsoft Graph invalidateAllRefreshTokens, or Microsoft Graph PowerShell, Revoke-MgUserSignInSession. Why is Azure AD B2C still accepting the old refresh token?
@@ -286,6 +292,9 @@ sections:
286292
1. Wait for 10 minutes.
287293
288294
1. Retrieve the `RefreshToken` again.
295+
296+
> [!TIP]
297+
> With Azure AD B2C [custom policy](custom-policy-overview.md), you can reduce the above mentioned skew time of 5 minutes (300000 milliseconds) by adjusting the value for InputParameter "TreatAsEqualIfWithinMillseconds" under claim transformation Id "AssertRefreshTokenIssuedLaterThanValidFromDate". This claim transformation can be found in the TrustFrameworkBase.xml file under latest custom policy [stater-pack](https://learn.microsoft.com/en-us/azure/active-directory-b2c/tutorial-create-user-flows?pivots=b2c-custom-policy#get-the-starter-pack).
289298
290299
- question: |
291300
I use multiple tabs in a web browser to sign in to multiple applications that I registered in the same Azure AD B2C tenant. When I try to perform a single sign-out, not all of the applications are signed out. Why does this happen?

articles/active-directory-b2c/javascript-and-page-layout.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ zone_pivot_groups: b2c-policy-type
2323

2424
[!INCLUDE [active-directory-b2c-choose-user-flow-or-custom-policy](../../includes/active-directory-b2c-choose-user-flow-or-custom-policy.md)]
2525

26-
With Azure Active Directory B2C (Azure AD B2C) [HTML templates](customize-ui-with-html.md), you can craft your users' identity experiences. Your HTML templates can contain only certain HTML tags and attributes. Basic HTML tags, such as &lt;b&gt;, &lt;i&gt;, &lt;u&gt;, &lt;h1&gt;, and &lt;hr&gt; are allowed. More advanced tags such as &lt;script&gt;, and &lt;iframe&gt; are removed for security reasons but the `<script>` tag should be added in the `<head>` tag.
26+
With Azure Active Directory B2C (Azure AD B2C) [HTML templates](customize-ui-with-html.md), you can craft your users' identity experiences. Your HTML templates can contain only certain HTML tags and attributes. Basic HTML tags, such as &lt;b&gt;, &lt;i&gt;, &lt;u&gt;, &lt;h1&gt;, and &lt;hr&gt; are allowed. More advanced tags such as &lt;script&gt;, and &lt;iframe&gt; are removed for security reasons but the `<script>` tag should be added in the `<head>` tag. From selfasserted page layout version 2.1.21 / unifiedssp version 2.1.10 / multifactor version 1.2.10 onwards B2C doesn't support adding scripts in `<body>` tag (as this can pose a risk for cross site scripting attack). Migrating existing scripts from `<body>` to `<head>` may at-times require rewriting existing scripts with mutation observers for proper working.
2727

2828
The `<script>` tag should be added in the `<head>` tag in two ways:
2929

articles/ai-services/computer-vision/concept-model-customization.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -80,8 +80,8 @@ The following table describes the limits on the scale of your custom model proje
8080
| Max # training images | 1,000,000 | 200,000 |
8181
| Max # evaluation images | 100,000 | 100,000 |
8282
| Min # training images per category | 2 | 2 |
83-
| Max # tags per image | multiclass: 1 | NA |
84-
| Max # regions per image | NA | 1,000 |
83+
| Max # tags per image | 1 | N/A |
84+
| Max # regions per image | N/A | 1,000 |
8585
| Max # categories | 2,500 | 1,000 |
8686
| Min # categories | 2 | 1 |
8787
| Max image size (Training) | 20 MB | 20 MB |

articles/ai-services/content-safety/overview.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,8 @@ This documentation contains the following article types:
3030

3131
The following are a few scenarios in which a software developer or team would require a content moderation service:
3232

33+
- User prompts submitted to a generative AI service.
34+
- Content produced by generative AI models.
3335
- Online marketplaces that moderate product catalogs and other user-generated content.
3436
- Gaming companies that moderate user-generated game artifacts and chat rooms.
3537
- Social messaging platforms that moderate images and text added by their users.

articles/ai-services/openai/how-to/gpt-with-vision.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ The following command shows the most basic way to use the GPT-4 Turbo with Visio
3030

3131
#### [REST](#tab/rest)
3232

33-
Send a POST request to `https://{RESOURCE_NAME}.openai.azure.com/openai/deployments/{DEPLOYMENT_NAME}/chat/completions?api-version=2023-12-01-preview` where
33+
Send a POST request to `https://{RESOURCE_NAME}.openai.azure.com/openai/deployments/{DEPLOYMENT_NAME}/chat/completions?api-version=2024-02-15-preview` where
3434

3535
- RESOURCE_NAME is the name of your Azure OpenAI resource
3636
- DEPLOYMENT_NAME is the name of your GPT-4 Turbo with Vision model deployment
@@ -88,12 +88,12 @@ The following is a sample request body. The format is the same as the chat compl
8888
api_base = '<your_azure_openai_endpoint>' # your endpoint should look like the following https://YOUR_RESOURCE_NAME.openai.azure.com/
8989
api_key="<your_azure_openai_key>"
9090
deployment_name = '<your_deployment_name>'
91-
api_version = '2023-12-01-preview' # this might change in the future
91+
api_version = '2024-02-15-preview' # this might change in the future
9292

9393
client = AzureOpenAI(
9494
api_key=api_key,
9595
api_version=api_version,
96-
base_url=f"{api_base}openai/deployments/{deployment_name}/extensions",
96+
base_url=f"{api_base}openai/deployments/{deployment_name}",
9797
)
9898
```
9999

@@ -265,7 +265,7 @@ The **object grounding** integration brings a new layer to data analysis and use
265265
266266
#### [REST](#tab/rest)
267267

268-
Send a POST request to `https://{RESOURCE_NAME}.openai.azure.com/openai/deployments/{DEPLOYMENT_NAME}/extensions/chat/completions?api-version=2023-12-01-preview` where
268+
Send a POST request to `https://{RESOURCE_NAME}.openai.azure.com/openai/deployments/{DEPLOYMENT_NAME}/chat/completions?api-version=2024-02-15-preview` where
269269

270270
- RESOURCE_NAME is the name of your Azure OpenAI resource
271271
- DEPLOYMENT_NAME is the name of your GPT-4 Turbo with Vision model deployment
@@ -577,7 +577,7 @@ To use a User assigned identity on your Azure AI Services resource, follow these
577577
578578
#### [REST](#tab/rest)
579579
580-
1. Prepare a POST request to `https://{RESOURCE_NAME}.openai.azure.com/openai/deployments/{DEPLOYMENT_NAME}/extensions/chat/completions?api-version=2023-12-01-preview` where
580+
1. Prepare a POST request to `https://{RESOURCE_NAME}.openai.azure.com/openai/deployments/{DEPLOYMENT_NAME}/chat/completions?api-version=2024-02-15-preview` where
581581
582582
- RESOURCE_NAME is the name of your Azure OpenAI resource
583583
- DEPLOYMENT_NAME is the name of your GPT-4 Vision model deployment

articles/ai-studio/how-to/deploy-models-mistral.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -64,8 +64,7 @@ Certain models in the model catalog can be deployed as a serverless API with pay
6464
- An [Azure AI Studio hub](../how-to/create-azure-ai-resource.md).
6565

6666
> [!IMPORTANT]
67-
> The serverless API model deployment offering for eligible models in the Mistral family is only available in hubs created in the **East US 2** and **Sweden Central** regions. For _Mistral Large_, the serverless API model deployment offering is also available in the **France Central** region.
68-
67+
> The serverless API model deployment offering for eligible models in the Mistral family is only available in hubs created in the **East US 2** and **Sweden Central** regions.
6968
- An [Azure AI Studio project](../how-to/create-projects.md).
7069
- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Studio. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, see [Role-based access control in Azure AI Studio](../concepts/rbac-ai-studio.md).
7170

@@ -94,7 +93,7 @@ To create a deployment:
9493

9594
:::image type="content" source="../media/deploy-monitor/mistral/mistral-large-deploy-pay-as-you-go.png" alt-text="A screenshot showing how to deploy a model as a serverless API." lightbox="../media/deploy-monitor/mistral/mistral-large-deploy-pay-as-you-go.png":::
9695

97-
1. Select the project in which you want to deploy your model. To deploy the Mistral model, your project must be in the *EastUS2* or *Sweden Central* region. For the Mistral Large model, you can also deploy in a project that's in the *France Central* region.
96+
1. Select the project in which you want to deploy your model. To deploy the Mistral model, your project must be in the *EastUS2* or *Sweden Central* region.
9897
1. In the deployment wizard, select the link to **Azure Marketplace Terms** to learn more about the terms of use.
9998
1. Select the **Pricing and terms** tab to learn about pricing for the selected model.
10099
1. Select the **Subscribe and Deploy** button. If this is your first time deploying the model in the project, you have to subscribe your project for the particular offering. This step requires that your account has the **Azure AI Developer role** permissions on the resource group, as listed in the prerequisites. Each project has its own subscription to the particular Azure Marketplace offering of the model, which allows you to control and monitor spending. Currently, you can have only one deployment for each model within a project.

articles/ai-studio/how-to/deploy-models-timegen-1.md

Lines changed: 36 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,42 @@ You can deploy TimeGEN-1 as a serverless API with pay-as-you-go billing. Nixtla
4747
- An [Azure AI Studio project](../how-to/create-projects.md).
4848
- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Studio. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, visit [Role-based access control in Azure AI Studio](../concepts/rbac-ai-studio.md).
4949

50+
### Pricing information
51+
52+
#### Estimate the number of tokens needed
53+
54+
Before you create a deployment, it's useful to estimate the number of tokens that you plan to use and be billed for.
55+
One token corresponds to one data point in your input dataset or output dataset.
56+
57+
Suppose you have the following input time series dataset:
58+
59+
| Unique_id | Timestamp | Target Variable | Exogenous Variable 1 | Exogenous Variable 2 |
60+
|:---------:|:-------------------:|:---------------:|:--------------------:|:--------------------:|
61+
| BE | 2016-10-22 00:00:00 | 70.00 | 49593.0 | 57253.0 |
62+
| BE | 2016-10-22 01:00:00 | 37.10 | 46073.0 | 51887.0 |
63+
64+
To determine the number of tokens, multiply the number of rows (in this example, two) and the number of columns used for forecasting—not counting the unique_id and timestamp columns (in this example, three) to get a total of six tokens.
65+
66+
Given the following output dataset:
67+
68+
| Unique_id | Timestamp | Forecasted Target Variable |
69+
|:---------:|:-------------------:|:--------------------------:|
70+
| BE | 2016-10-22 02:00:00 | 46.57 |
71+
| BE | 2016-10-22 03:00:00 | 48.57 |
72+
73+
You can also determine the number of tokens by counting the number of data points returned after data forecasting. In this example, the number of tokens is two.
74+
75+
#### Estimate the pricing
76+
77+
There are four pricing meters, as described in the following table:
78+
79+
| Pricing Meter | Description |
80+
|-----------------------------------------------|-----------------------------------------------------------------------------------------|
81+
| paygo-inference-input-tokens | Costs associated with the tokens used as input for inference when *finetune_steps* = 0 |
82+
| paygo-inference-output-tokens | Costs associated with the tokens used as output for inference when *finetune_steps* = 0 |
83+
| paygo-finetuned-model-inference-input-tokens | Costs associated with the tokens used as input for inference when *finetune_steps* > 0 |
84+
| paygo-finetuned-model-inference-output-tokens | Costs associated with the tokens used as output for inference when *finetune_steps* > 0 |
85+
5086
### Create a new deployment
5187

5288
These steps demonstrate the deployment of TimeGEN-1. To create a deployment:

articles/ai-studio/reference/reference-model-inference-api.md

Lines changed: 28 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -136,7 +136,7 @@ Content-Type: application/json
136136

137137
The Azure AI Model Inference API specifies a set of modalities and parameters that models can subscribe to. However, some models may have further capabilities that the ones the API indicates. On those cases, the API allows the developer to pass them as extra parameters in the payload.
138138

139-
By setting a header `extra-parameters: allow`, the API will attempt to pass any unknown parameter directly to the underlying model. If the model can handle that parameter, the request completes.
139+
By setting a header `extra-parameters: pass-through`, the API will attempt to pass any unknown parameter directly to the underlying model. If the model can handle that parameter, the request completes.
140140

141141
The following example shows a request passing the parameter `safe_prompt` supported by Mistral-Large, which isn't specified in the Azure AI Model Inference API:
142142

@@ -163,6 +163,7 @@ var messages = [
163163
];
164164

165165
var response = await client.path("/chat/completions").post({
166+
"extra-parameters": "pass-through",
166167
body: {
167168
messages: messages,
168169
safe_mode: true
@@ -178,7 +179,7 @@ __Request__
178179
POST /chat/completions?api-version=2024-04-01-preview
179180
Authorization: Bearer <bearer-token>
180181
Content-Type: application/json
181-
extra-parameters: allow
182+
extra-parameters: pass-through
182183
```
183184

184185
```JSON
@@ -203,7 +204,7 @@ extra-parameters: allow
203204
---
204205

205206
> [!TIP]
206-
> Alternatively, you can set `extra-parameters: drop` to drop any unknown parameter in the request. Use this capability in case you happen to be sending requests with extra parameters that you know the model won't support but you want the request to completes anyway. A typical example of this is indicating `seed` parameter.
207+
> The default value for `extra-parameters` is `error` which returns an error if an extra parameter is indicated in the payload. Alternatively, you can set `extra-parameters: ignore` to drop any unknown parameter in the request. Use this capability in case you happen to be sending requests with extra parameters that you know the model won't support but you want the request to completes anyway. A typical example of this is indicating `seed` parameter.
207208
208209
### Models with disparate set of capabilities
209210

@@ -426,3 +427,27 @@ __Response__
426427
## Getting started
427428

428429
The Azure AI Model Inference API is currently supported in certain models deployed as [Serverless API endpoints](../how-to/deploy-models-serverless.md) and Managed Online Endpoints. Deploy any of the [supported models](#availability) and use the exact same code to consume their predictions.
430+
431+
# [Python](#tab/python)
432+
433+
The client library `azure-ai-inference` does inference, including chat completions, for AI models deployed by Azure AI Studio and Azure Machine Learning Studio. It supports Serverless API endpoints and Managed Compute endpoints (formerly known as Managed Online Endpoints).
434+
435+
Explore our [samples](https://github.com/Azure/azure-sdk-for-python/tree/main/sdk/ai/azure-ai-inference/samples) and read the [API reference documentation](https://aka.ms/azsdk/azure-ai-inference/python/reference) to get yourself started.
436+
437+
# [JavaScript](#tab/javascript)
438+
439+
The client library `@azure-rest/ai-inference` does inference, including chat completions, for AI models deployed by Azure AI Studio and Azure Machine Learning Studio. It supports Serverless API endpoints and Managed Compute endpoints (formerly known as Managed Online Endpoints).
440+
441+
Explore our [samples](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/ai/ai-inference-rest/samples) and read the [API reference documentation](https://aka.ms/AAp1kxa) to get yourself started.
442+
443+
# [REST](#tab/rest)
444+
445+
Explore the reference section of the Azure AI model inference API to see parameters and options to consume models, including chat completions models, deployed by Azure AI Studio and Azure Machine Learning Studio. It supports Serverless API endpoints and Managed Compute endpoints (formerly known as Managed Online Endpoints).
446+
447+
* [Get info](reference-model-inference-info.md): Returns the information about the model deployed under the endpoint.
448+
* [Text embeddings](reference-model-inference-embeddings.md): Creates an embedding vector representing the input text.
449+
* [Text completions](reference-model-inference-completions.md): Creates a completion for the provided prompt and parameters.
450+
* [Chat completions](reference-model-inference-chat-completions.md): Creates a model response for the given chat conversation.
451+
* [Image embeddings](reference-model-inference-images-embeddings.md): Creates an embedding vector representing the input text and image.
452+
453+
---

articles/ai-studio/reference/reference-model-inference-chat-completions.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ POST /chat/completions?api-version=2024-04-01-preview
3535

3636
| Name | Required | Type | Description |
3737
| --- | --- | --- | --- |
38-
| extra-parameters | | string | The behavior of the API when extra parameters are indicated in the payload. Using `allow` makes the API to pass the parameter to the underlying model. Use this value when you want to pass parameters that you know the underlying model can support. Using `drop` makes the API to drop any unsupported parameter. Use this value when you need to use the same payload across different models, but one of the extra parameters may make a model to error out if not supported. Using `error` makes the API to reject any extra parameter in the payload. Only parameters specified in this API can be indicated, or a 400 error is returned. |
38+
| extra-parameters | | string | The behavior of the API when extra parameters are indicated in the payload. Using `pass-through` makes the API to pass the parameter to the underlying model. Use this value when you want to pass parameters that you know the underlying model can support. Using `ignore` makes the API to drop any unsupported parameter. Use this value when you need to use the same payload across different models, but one of the extra parameters may make a model to error out if not supported. Using `error` makes the API to reject any extra parameter in the payload. Only parameters specified in this API can be indicated, or a 400 error is returned. |
3939
| azureml-model-deployment | | string | Name of the deployment you want to route the request to. Supported for endpoints that support multiple deployments. |
4040

4141
## Request Body

0 commit comments

Comments
 (0)