You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-services/openai/overview.md
+20-7Lines changed: 20 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,15 +27,28 @@ Azure OpenAI Service provides REST API access to OpenAI's powerful language mode
27
27
| Managed Identity| Yes, via Microsoft Entra ID |
28
28
| UI experience |[Azure portal](https://portal.azure.com) for account & resource management, <br> [Azure AI Studio](https://ai.azure.com) for model exploration and fine-tuning |
29
29
| Model regional availability |[Model availability](./concepts/models.md)|
30
-
| Content filtering | Prompts and completions are evaluated against our content policy with automated systems. High severity content will be filtered. |
30
+
| Content filtering | Prompts and completions are evaluated against our content policy with automated systems. High severity content is filtered. |
31
31
32
32
## Responsible AI
33
33
34
34
At Microsoft, we're committed to the advancement of AI driven by principles that put people first. Generative models such as the ones available in Azure OpenAI have significant potential benefits, but without careful design and thoughtful mitigations, such models have the potential to generate incorrect or even harmful content. Microsoft has made significant investments to help guard against abuse and unintended harm, which includes incorporating Microsoft’s <a href="https://www.microsoft.com/ai/responsible-ai?activetab=pivot1:primaryr6" target="_blank">principles for responsible AI use</a>, adopting a [Code of Conduct](/legal/cognitive-services/openai/code-of-conduct?context=/azure/ai-services/openai/context/context) for use of the service, building [content filters](/azure/ai-services/content-safety/overview) to support customers, and providing responsible AI [information and guidance](/legal/cognitive-services/openai/transparency-note?context=%2Fazure%2Fai-services%2Fopenai%2Fcontext%2Fcontext&tabs=image) that customers should consider when using Azure OpenAI.
35
35
36
-
## How do I get access to Azure OpenAI?
36
+
## Get started with Azure OpenAI Service
37
37
38
-
A Limited Access registration form is not required to access most Azure OpenAI models. Learn more on the [Azure OpenAI Limited Access page](/legal/cognitive-services/openai/limited-access?context=/azure/ai-services/openai/context/context).
38
+
To get started with Azure OpenAI Service, you need to create an Azure OpenAI Service resource in your Azure subscription.
39
+
40
+
Start with the [Create and deploy an Azure OpenAI Service resource](./how-to/create-resource.md) guide.
41
+
1. You can create a resource via Azure portal, Azure CLI, or Azure PowerShell.
42
+
1. When you have an Azure OpenAI Service resource, you can deploy a model such as GPT-4o.
43
+
1. When you have a deployed model, you can:
44
+
45
+
- Try out the Azure AI Studio playgrounds to explore the capabilities of the models.
46
+
- You can also just start making API calls to the service using the REST API or SDKs.
47
+
48
+
For example, you can try [real-time audio](./realtime-audio-quickstart.md) and [assistants](./assistants-quickstart.md) in the playgrounds or via code.
49
+
50
+
> [!NOTE]
51
+
> A Limited Access registration form is required to access some Azure OpenAI Service models or features. Learn more on the [Azure OpenAI Limited Access page](/legal/cognitive-services/openai/limited-access?context=/azure/ai-services/openai/context/context).
39
52
40
53
## Comparing Azure OpenAI and OpenAI
41
54
@@ -47,7 +60,7 @@ With Azure OpenAI, customers get the security capabilities of Microsoft Azure wh
47
60
48
61
### Prompts & completions
49
62
50
-
The completions endpoint is the core component of the API service. This API provides access to the model's text-in, text-out interface. Users simply need to provide an input **prompt** containing the English text command, and the model will generate a text **completion**.
63
+
The completions endpoint is the core component of the API service. This API provides access to the model's text-in, text-out interface. Users simply need to provide an input **prompt** containing the English text command, and the model generates a text **completion**.
51
64
52
65
Here's an example of a simple prompt and completion:
53
66
@@ -70,7 +83,7 @@ Here's an example of a simple prompt and completion:
70
83
71
84
Azure OpenAI processes text by breaking it down into tokens. Tokens can be words or just chunks of characters. For example, the word “hamburger” gets broken up into the tokens “ham”, “bur” and “ger”, while a short and common word like “pear” is a single token. Many tokens start with a whitespace, for example “ hello” and “ bye”.
72
85
73
-
The total number of tokens processed in a given request depends on the length of your input, output and request parameters. The quantity of tokens being processed will also affect your response latency and throughput for the models.
86
+
The total number of tokens processed in a given request depends on the length of your input, output, and request parameters. The quantity of tokens being processed will also affect your response latency and throughput for the models.
74
87
75
88
#### Image tokens
76
89
@@ -106,9 +119,9 @@ Once you create an Azure OpenAI Resource, you must deploy a model before you can
106
119
107
120
### Prompt engineering
108
121
109
-
The GPT-3, GPT-3.5 and GPT-4 models from OpenAI are prompt-based. With prompt-based models, the user interacts with the model by entering a text prompt, to which the model responds with a text completion. This completion is the model’s continuation of the input text.
122
+
The GPT-3, GPT-3.5, and GPT-4 models from OpenAI are prompt-based. With prompt-based models, the user interacts with the model by entering a text prompt, to which the model responds with a text completion. This completion is the model’s continuation of the input text.
110
123
111
-
While these models are extremely powerful, their behavior is also very sensitive to the prompt. This makes [prompt engineering](./concepts/prompt-engineering.md) an important skill to develop.
124
+
While these models are powerful, their behavior is also sensitive to the prompt. This makes [prompt engineering](./concepts/prompt-engineering.md) an important skill to develop.
112
125
113
126
Prompt construction can be difficult. In practice, the prompt acts to configure the model weights to complete the desired task, but it's more of an art than a science, often requiring experience and intuition to craft a successful prompt.
Copy file name to clipboardExpand all lines: articles/machine-learning/concept-compute-target.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -92,7 +92,7 @@ When created, these compute resources are automatically part of your workspace,
92
92
> [!NOTE]
93
93
> To avoid charges when the compute is idle:
94
94
> * For a compute *cluster*, make sure the minimum number of nodes is set to 0, or use [serverless compute](./how-to-use-serverless-compute.md).
95
-
> * For a compute *instance*, [enable idle shutdown](how-to-create-compute-instance.md#configure-idle-shutdown).
95
+
> * For a compute *instance*, [enable idle shutdown](how-to-create-compute-instance.md#configure-idle-shutdown). While stopping the compute instance stops the billing for compute hours, you'll still be billed for disk, public IP, and standard load balancer.
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-create-compute-instance.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -62,7 +62,7 @@ Choose the tab for the environment you're using for other prerequisites.
62
62
63
63
Creating a compute instance is a one time process for your workspace. You can reuse the compute as a development workstation or as a compute target for training. You can have multiple compute instances attached to your workspace.
64
64
65
-
The dedicated cores per region per VM family quota and total regional quota, which applies to compute instance creation, is unified and shared with Azure Machine Learning training compute cluster quota. Stopping the compute instance doesn't release quota to ensure you are able to restart the compute instance. It isn't possible to change the virtual machine size of compute instance once it's created.
65
+
The dedicated cores per region per VM family quota and total regional quota, which applies to compute instance creation, is unified and shared with Azure Machine Learning training compute cluster quota. Stopping the compute instance doesn't release quota to ensure you are able to restart the compute instance. A quota is a credit limit on Azure resources, not a capacity guarantee, Restarting a compute instance is still dependent on the available capacity of the region. If there is a capacity crunch in the region for the SKU, you may not be able to restart the compute instance. It isn't possible to change the virtual machine size of compute instance once it's created.
66
66
67
67
The fastest way to create a compute instance is to follow the [Create resources you need to get started](quickstart-create-resources.md).
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-manage-compute-instance.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -64,7 +64,7 @@ You can [enable automatic shutdown](how-to-create-compute-instance.md#configure-
64
64
You can also [create a schedule](how-to-create-compute-instance.md#schedule-automatic-start-and-stop) for the compute instance to automatically start and stop based on a time and day of week.
65
65
66
66
> [!TIP]
67
-
> The compute instance has 120GB OS disk. If you run out of disk space, [use the terminal](how-to-access-terminal.md) to clear at least 1-2 GB before you stop or restart the compute instance. Please do not stop the compute instance by issuing sudo shutdown from the terminal. The temp disk size on compute instance depends on the VM size chosen and is mounted on /mnt.
67
+
> The compute instance has 120GB OS disk. If you run out of disk space, [use the terminal](how-to-access-terminal.md) to clear at least 5 GB before you stop or restart the compute instance. Please do not stop the compute instance by issuing sudo shutdown from the terminal. The temp disk size on compute instance depends on the VM size chosen and is mounted on /mnt.
0 commit comments