@@ -52,20 +52,18 @@ items:
52
52
items :
53
53
- name : Deploying models in Azure AI Foundry
54
54
href : concepts/deployments-overview.md
55
- - name : Serverless API
56
- items :
57
- - name : Deploy models as serverless API
58
- href : how-to/deploy-models-serverless.md
59
- - name : Consume serverless API models from a different project or hub
60
- href : how-to/deploy-models-serverless-connect.md
61
- - name : Model and region availability for Serverless API deployments
62
- href : how-to/deploy-models-serverless-availability.md
63
- - name : Content safety for models deployed with serverless APIs
64
- href : concepts/model-catalog-content-safety.md
65
- - name : Managed compute
55
+ - name : Azure OpenAI Service
66
56
items :
67
- - name : Deploy models via managed compute
68
- href : how-to/deploy-models-managed.md
57
+ - name : Azure OpenAI in Azure AI Foundry
58
+ href : azure-openai-in-ai-foundry.md
59
+ - name : Use Azure OpenAI Service in Azure AI Foundry portal
60
+ href : ai-services/how-to/connect-azure-openai.md
61
+ - name : Deploy Azure OpenAI models
62
+ href : how-to/deploy-models-openai.md
63
+ - name : Azure OpenAI Service quotas and limits
64
+ href : ../ai-services/openai/quotas-limits.md?context=/azure/ai-foundry/context/context
65
+ - name : Troubleshoot deployments and monitoring
66
+ href : how-to/troubleshoot-deploy-and-monitor.md
69
67
- name : Azure AI model inference
70
68
items :
71
69
- name : What is Azure AI model inference?
@@ -82,18 +80,20 @@ items:
82
80
href : ../ai-foundry/model-inference/how-to/inference.md?context=/azure/ai-foundry/context/context
83
81
- name : Azure AI model inference quotas and limits
84
82
href : ../ai-foundry/model-inference/quotas-limits.md?context=/azure/ai-foundry/context/context
85
- - name : Azure OpenAI Service
83
+ - name : Serverless API
86
84
items :
87
- - name : Azure OpenAI in Azure AI Foundry
88
- href : azure-openai-in-ai-foundry.md
89
- - name : Use Azure OpenAI Service in Azure AI Foundry portal
90
- href : ai-services/how-to/connect-azure-openai.md
91
- - name : Deploy Azure OpenAI models
92
- href : how-to/deploy-models-openai.md
93
- - name : Azure OpenAI Service quotas and limits
94
- href : ../ai-services/openai/quotas-limits.md?context=/azure/ai-foundry/context/context
95
- - name : Troubleshoot deployments and monitoring
96
- href : how-to/troubleshoot-deploy-and-monitor.md
85
+ - name : Deploy models as serverless API
86
+ href : how-to/deploy-models-serverless.md
87
+ - name : Consume serverless API models from a different project or hub
88
+ href : how-to/deploy-models-serverless-connect.md
89
+ - name : Model and region availability for Serverless API deployments
90
+ href : how-to/deploy-models-serverless-availability.md
91
+ - name : Content safety for models deployed with serverless APIs
92
+ href : concepts/model-catalog-content-safety.md
93
+ - name : Managed compute
94
+ items :
95
+ - name : Deploy models via managed compute
96
+ href : how-to/deploy-models-managed.md
97
97
- name : Work with models from the model catalog
98
98
items :
99
99
- name : Featured models in the model catalog
0 commit comments