Skip to content

Commit f019439

Browse files
committed
Learn Editor: Update working-with-models.md
1 parent 42289d2 commit f019439

File tree

1 file changed

+58
-1
lines changed

1 file changed

+58
-1
lines changed

articles/ai-services/openai/how-to/working-with-models.md

Lines changed: 58 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -292,7 +292,7 @@ Provisioned deployments support distinct model management practices. Provisioned
292292
- For multi-deployment migrations, validate that you have sufficient quota to support multiple deployments simultaneously. For more information on how to validate quota for each provisioned deployment type, review the [provisioned quota documentation](../concepts/provisioned-throughput.md#quota).
293293

294294
### In-place migrations for provisioned deployments
295-
In-place migrations allow you to maintain the same provisioned deployment name and size while changing the model version or model family assigned to that deployment. With in-place migrations, Azure OpenAI Service takes care of migrating any existing traffic between model versions or model families throughout the migration over a 20-30 minute window. Throughout the migration window, your provisioned deployment will display an "updating" provisioned state. Once the in-place migration is complete, the provisioned state will be updated to "succeeded", indicating that all traffic has been migrated over to the target model version or model family.
295+
In-place migrations allow you to maintain the same provisioned deployment name and size while changing the model version or model family assigned to that deployment. With in-place migrations, Azure OpenAI Service takes care of migrating any existing traffic between model versions or model families throughout the migration over a 20-30 minute window. Throughout the migration window, your provisioned deployment will display an "updating" provisioned state. You can continue to use your provisioned deployment as you normally would. Once the in-place migration is complete, the provisioned state will be updated to "succeeded", indicating that all traffic has been migrated over to the target model version or model family.
296296

297297
#### In-place migration: model version update
298298
In-place migrations that target updating an existing provisioned deployment to a new model version within the same model family are supported through Azure AI Foundry, REST API, and Azure CLI. To perform an in-place migration targeting a model version update within Azure AI Foundry, select **Deployments** > under the deployment name column select the deployment name of the provisioned deployment you would like to migrate.
@@ -302,6 +302,63 @@ Selecting a deployment name opens the **Properties** for the model deployment. F
302302
![Screenshot of update deployment dialogue box with the model version field selector opened to show model version options available for selection.](media/working-with-models/provisioned-deployment-model-version-update.png)
303303

304304
#### In-place migration: model family change
305+
In-place migration that target updating an existing provisioned deployment to a new model family are supported through REST API and Azure CLI. To perform an in-place migration targeting a model family change, use the example request below as a guide. In the request, you will need to update the model name and model version for the target model you are migrating to.
306+
307+
```Bash
308+
curl -X PUT https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/resource-group-temp/providers/Microsoft.CognitiveServices/accounts/docs-openai-test-001/deployments/gpt-4o-ptu-deployment?api-version=2024-10-01 \
309+
-H "Content-Type: application/json" \
310+
-H 'Authorization: Bearer YOUR_AUTH_TOKEN' \
311+
-d '{"sku":{"name":"GlobalProvisionedManaged","capacity":100},"properties": {"model": {"format": "OpenAI","name": "gpt-4o-mini","version": "2024-07-18"}}}'
312+
```
313+
#### Example response
314+
315+
```json
316+
{
317+
"id": "/subscriptions/{subscription-id}/resourceGroups/resource-group-temp/providers/Microsoft.CognitiveServices/accounts/docs-openai-test-001/deployments/gpt-35-turbo",
318+
"type": "Microsoft.CognitiveServices/accounts/deployments",
319+
"name": "gpt-4o-ptu-deployment",
320+
"sku": {
321+
"name": "GlobalProvisionedManaged",
322+
"capacity": 100
323+
},
324+
"properties": {
325+
"model": {
326+
"format": "OpenAI",
327+
"name": "gpt-4o-mini",
328+
"version": "2024-07-18"
329+
},
330+
"versionUpgradeOption": "OnceCurrentVersionExpired",
331+
"currentCapacity": 100
332+
"capabilities": {
333+
"area": "EUR",
334+
"chatCompletion": "true"
335+
"jsonObjectResponse": "true",
336+
"maxContextToken": "128000",
337+
"maxOutputToken": "16834",
338+
"assistants": "true"
339+
},
340+
"provisioningState": "Updating",
341+
"rateLimits": [
342+
{
343+
"key": "request",
344+
"renewalPeriod": 10,
345+
"count": 120
346+
}
347+
]
348+
},
349+
"systemData": {
350+
"createdBy": "[email protected]",
351+
"createdByType": "User",
352+
"createdAt": "2025-01-28T02:57:15.8951706Z",
353+
"lastModifiedBy": "[email protected]",
354+
"lastModifiedByType": "User",
355+
"lastModifiedAt": "2025-01-29T15:35:53.082912Z"
356+
},
357+
"etag": "\"GUID\""
358+
}
359+
360+
> [!NOTE]
361+
> There are multiple ways to generate an authorization token. The easiest method for initial testing is to launch the Cloud Shell from the [Azure portal](https://portal.azure.com). Then run [`az account get-access-token`](/cli/azure/account?view=azure-cli-latest#az-account-get-access-token&preserve-view=true). You can use this token as your temporary authorization token for API testing.
305362

306363
### Multi-deployment migrations for provisioned deployments
307364

0 commit comments

Comments
 (0)