You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> There is no additional [pricing](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) or [quota](../quotas-limits.md) for using Assistants unless you use the [code interpreter](../how-to/code-interpreter.md) tool.
30
+
28
31
Assistant API is built on the same capabilities that power OpenAI’s GPT product. Some possible use cases range from AI-powered product recommender, sales analyst app, coding assistant, employee Q&A chatbot, and more. Start building on the no-code Assistants playground on the Azure OpenAI Studio or start building with the API.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/faq.yml
+54-3Lines changed: 54 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -118,7 +118,7 @@ sections:
118
118
answer:
119
119
If the service performs processing, you will be charged even if the status code is not successful (not 200).
120
120
Common examples of this are, a 400 error due to a content filter or input limit, or a 408 error due to a timeout. Charges will also occur when a `status 200` is received with a `finish_reason` of `content_filter`.
121
-
In this case the prompt did not have any issues, but the completion generated by the model was detected to violate the content filtering rules which results in the completion being filtered.
121
+
In this case the prompt did not have any issues, but the completion generated by the model was detected to violate the content filtering rules, which result in the completion being filtered.
122
122
123
123
If the service doesn't perform processing, you won't be charged.
124
124
For example, a 401 error due to authentication or a 429 error due to exceeding the Rate Limit.
@@ -228,7 +228,58 @@ sections:
228
228
What are the known limitations of GPT-4 Turbo with Vision?
229
229
answer: |
230
230
See the [limitations](./concepts/gpt-with-vision.md#limitations) section of the GPT-4 Turbo with Vision concepts guide.
231
+
- name: Assistants
232
+
questions:
233
+
- question: |
234
+
Do you store any data used in the Assistants API?
235
+
answer: |
236
+
Yes. Unlike Chat Completions API, Azure OpenAI Assistants is a stateful API, meaning it retains data. There are two types of data stored in the Assistants API:
237
+
* Stateful entities: Threads, messages, and runs created during Assistants use.
238
+
* Files: Uploaded during Assistants setup or as part of a message.
239
+
- question: |
240
+
Where is this data stored?
241
+
answer: |
242
+
Data is stored in a secure, Microsoft-managed storage account that is logically separated.
243
+
- question: |
244
+
How long is this data stored?
245
+
answer: |
246
+
All used data persists in this system unless you explicitly delete this data. Use the [delete function](./assistants-reference-threads.md) with the thread ID of the thread you want to delete. Clearing the Run in the Assistants Playground does not delete threads, however deleting them using delete function will not list them in the thread page.
247
+
- question: |
248
+
Can I bring my own data store to use with Assistants?
249
+
answer: |
250
+
No. Currently Assistants supports only local files uploaded to the Assistants-managed storage. You cannot use your private storage account with Assistants.
251
+
- question: |
252
+
Is my data used by Microsoft for training models?
253
+
answer: |
254
+
No. Data is not used for Microsoft not used for training models. See the [Responsible AI documentation](/legal/cognitive-services/openai/data-privacy?context=%2Fazure%2Fai-services%2Fopenai%2Fcontext%2Fcontext) for more information.
255
+
- question: |
256
+
Where is data stored geographically?
257
+
answer: |
258
+
Azure OpenAI Assistants endpoints are regional, and data is stored in the same region as the endpoint. For more information, see the [Azure data residency documentation](https://azure.microsoft.com/explore/global-infrastructure/data-residency/#overview).
259
+
- question: |
260
+
How am I charged for Assistants?
261
+
answer: |
262
+
Currently, when you use Assistants API, you're billed for the following:
263
+
- Inference cost (input and output) of the base model you're using for each Assistant (for example gpt-4-0125). If you've created multiple Assistants, you will be charged for the base model attached to each Assistant.
264
+
- If you've enabled the Code Interpreter tool. For example if your assistant calls Code Interpreter simultaneously in two different threads, this would create two Code Interpreter sessions, each of which would be charged. Each session is active by default for one hour, which means that you would only pay this fee once if your user keeps giving instructions to Code Interpreter in the same thread for up to one hour.
231
265
266
+
For more information, see the [pricing page](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/).
267
+
- question: |
268
+
Is there any additional pricing or quota for using Assistants?
269
+
answer: |
270
+
No. All [quotas](./quotas-limits.md) apply to using models with Assistants.
271
+
- question: |
272
+
Does the Assistants API support non-Azure OpenAI models?
273
+
answer: |
274
+
Assistants API only supports Azure OpenAI models.
275
+
- question: |
276
+
Is the Assistants API generally available?
277
+
answer: |
278
+
The Assistants API is currently in public preview. Stay informed of our latest product updates by regularly visiting our [What's New](./whats-new.md) page.
279
+
- question: |
280
+
What are some examples or other resources I can use to learn about Assistants?
281
+
answer: |
282
+
See the [Conceptual](./concepts/assistants.md), [quickstart](./assistants-quickstart.md), [how-to](./how-to/assistant.md) articles for information on getting started and using Assistants. You can also check out Azure OpenAI Assistants code samples on [GitHub](https://github.com/Azure-Samples/azureai-samples/tree/main/scenarios/Assistants).
232
283
- name: Web app
233
284
questions:
234
285
- question: |
@@ -260,7 +311,7 @@ sections:
260
311
- question: |
261
312
How can I customize or automate the index creation process?
262
313
answer:
263
-
You can prepare the index yourself using a [script provided on GitHub](https://go.microsoft.com/fwlink/?linkid=2244395). Using this script will create an Azure AI Search index with all the information needed to better leverage your data, with your documents broken down into manageable chunks. Please see the README file with the data preparation code for details on how to run it.
314
+
You can prepare the index yourself using a [script provided on GitHub](https://go.microsoft.com/fwlink/?linkid=2244395). Using this script will create an Azure AI Search index with all the information needed to better use your data, with your documents broken down into manageable chunks. See the README file with the data preparation code for details on how to run it.
264
315
- question: |
265
316
How can I update my index?
266
317
answer:
@@ -289,7 +340,7 @@ sections:
289
340
If Semantic Search is enabled for my Azure AI Search resource, will it be automatically applied to Azure OpenAI on your data in the Azure OpenAI Studio?
290
341
answer:
291
342
When you select "Azure AI Search" as the data source, you can choose to apply semantic search.
292
-
If you select "Azure Blob Container" or "Upload files" as the data source, you can create the index as usual. Afterwards you would re-ingest the data using the "Azure AI Search" option to select the same index and apply Semantic Search. You will then be ready to chat on your data with semantic search applied.
343
+
If you select "Azure Blob Container" or "Upload files" as the data source, you can create the index as usual. Afterwards you would reingest the data using the "Azure AI Search" option to select the same index and apply Semantic Search. You will then be ready to chat on your data with semantic search applied.
293
344
- question: |
294
345
How can I add vector embeddings when indexing my data?
0 commit comments