Skip to content

Commit a186e65

Browse files
authored
Merge pull request #896 from MicrosoftDocs/main
10/18 11:00 AM IST Publish
2 parents 6e9a4a2 + d213900 commit a186e65

15 files changed

+65
-81
lines changed

articles/ai-services/openai/how-to/batch.md

Lines changed: 0 additions & 65 deletions
Original file line numberDiff line numberDiff line change
@@ -157,71 +157,6 @@ Yes. Similar to other deployment types, you can create content filters and assoc
157157

158158
Yes, from the quota page in the Studio UI. Default quota allocation can be found in the [quota and limits article](../quotas-limits.md#global-batch-quota).
159159

160-
### How do I tell how many tokens my batch request contains, and how many tokens are available as quota?
161-
162-
The `2024-10-01-preview` REST API adds two new response headers:
163-
164-
* `deployment-enqueued-tokens` - A approximate token count for your jsonl file calculated immediately after the batch request is submitted. This value represents an estimate based on the number of characters and is not the true token count.
165-
* `deployment-maximum-enqueued-tokens` The total available enqueued tokens available for this global batch model deployment.
166-
167-
These response headers are only available when making a POST request to begin batch processing of a file with the REST API. The language specific client libraries do not currently return these new response headers. To return all response headers you can add `-i` to the standard REST request.
168-
169-
```http
170-
curl -i -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-01-preview \
171-
-H "api-key: $AZURE_OPENAI_API_KEY" \
172-
-H "Content-Type: application/json" \
173-
-d '{
174-
"input_file_id": "file-abc123",
175-
"endpoint": "/chat/completions",
176-
"completion_window": "24h"
177-
}'
178-
```
179-
180-
```output
181-
HTTP/1.1 200 OK
182-
Content-Length: 619
183-
Content-Type: application/json; charset=utf-8
184-
Vary: Accept-Encoding
185-
Request-Context: appId=
186-
x-ms-response-type: standard
187-
deployment-enqueued-tokens: 139
188-
deployment-maximum-enqueued-tokens: 740000
189-
Strict-Transport-Security: max-age=31536000; includeSubDomains; preload
190-
X-Content-Type-Options: nosniff
191-
x-aml-cluster: vienna-swedencentral-01
192-
x-request-time: 2.125
193-
apim-request-id: c8bf4351-c6f5-4bfe-9a79-ef3720eca8af
194-
x-ms-region: Sweden Central
195-
Date: Thu, 17 Oct 2024 01:45:45 GMT
196-
197-
{
198-
"cancelled_at": null,
199-
"cancelling_at": null,
200-
"completed_at": null,
201-
"completion_window": "24h",
202-
"created_at": 1729129545,
203-
"error_file_id": null,
204-
"expired_at": null,
205-
"expires_at": 1729215945,
206-
"failed_at": null,
207-
"finalizing_at": null,
208-
"id": "batch_c8dd49a7-c808-4575-9957-b188cd0dd642",
209-
"in_progress_at": null,
210-
"input_file_id": "file-f89384af0082485da43cb26b49dc25ce",
211-
"errors": null,
212-
"metadata": null,
213-
"object": "batch",
214-
"output_file_id": null,
215-
"request_counts": {
216-
"total": 0,
217-
"completed": 0,
218-
"failed": 0
219-
},
220-
"status": "validating",
221-
"endpoint": "/chat/completions"
222-
}
223-
```
224-
225160
### What happens if the API doesn't complete my request within the 24 hour time frame?
226161

227162
We aim to process these requests within 24 hours; we don't expire the jobs that take longer. You can cancel the job anytime. When you cancel the job, any remaining work is cancelled and any already completed work is returned. You'll be charged for any completed work.

articles/ai-services/openai/how-to/code-interpreter.md

Lines changed: 12 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -6,16 +6,16 @@ services: cognitive-services
66
manager: nitinme
77
ms.service: azure-ai-openai
88
ms.topic: how-to
9-
ms.date: 05/20/2024
10-
author: mrbullwinkle
11-
ms.author: mbullwin
9+
ms.date: 10/15/2024
10+
author: aahill
11+
ms.author: aahi
1212
recommendations: false
1313

1414
---
1515

1616
# Azure OpenAI Assistants Code Interpreter (Preview)
1717

18-
Code Interpreter allows the Assistants API to write and run Python code in a sandboxed execution environment. With Code Interpreter enabled, your Assistant can run code iteratively to solve more challenging code, math, and data analysis problems. When your Assistant writes code that fails to run, it can iterate on this code by modifying and running different code until the code execution succeeds.
18+
Code Interpreter allows the Assistants API to write and run Python code in a sandboxed execution environment. With Code Interpreter enabled, your Assistant can run code iteratively to solve more challenging code, math, and data analysis problems. When your Assistant writes code that fails to run, it can iterate on this code by modifying and running different code until the code execution succeeds.
1919

2020
> [!IMPORTANT]
2121
> Code Interpreter has [additional charges](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/) beyond the token based fees for Azure OpenAI usage. If your Assistant calls Code Interpreter simultaneously in two different threads, two code interpreter sessions are created. Each session is active by default for one hour.
@@ -28,7 +28,7 @@ Code Interpreter allows the Assistants API to write and run Python code in a san
2828

2929
The [models page](../concepts/models.md#assistants-preview) contains the most up-to-date information on regions/models where Assistants and code interpreter are supported.
3030

31-
We recommend using assistants with the latest models to take advantage of the new features, as well as the larger context windows, and more up-to-date training data.
31+
We recommend using assistants with the latest models to take advantage of the new features, larger context windows, and more up-to-date training data.
3232

3333
### API Versions
3434

@@ -69,7 +69,7 @@ We recommend using assistants with the latest models to take advantage of the ne
6969

7070
### File upload API reference
7171

72-
Assistants use the [same API for file upload as fine-tuning](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-02-15-preview&tabs=HTTP&preserve-view=true). When uploading a file you have to specify an appropriate value for the [purpose parameter](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-02-15-preview&tabs=HTTP&preserve-view=true#purpose).
72+
Assistants use the [same API for file upload as fine-tuning](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-02-15-preview&tabs=HTTP&preserve-view=true). When uploading a file, you have to specify an appropriate value for the [purpose parameter](/rest/api/azureopenai/files/upload?view=rest-azureopenai-2024-02-15-preview&tabs=HTTP&preserve-view=true#purpose).
7373

7474
## Enable Code Interpreter
7575

@@ -136,7 +136,7 @@ assistant = client.beta.assistants.create(
136136
instructions="You are an AI assistant that can write code to help answer math questions.",
137137
model="gpt-4-1106-preview",
138138
tools=[{"type": "code_interpreter"}],
139-
file_ids=[file.id]
139+
tool_resources={"code interpreter":{"file_ids":[file.id]}}
140140
)
141141
```
142142

@@ -161,7 +161,11 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/assistants?api-version=2
161161
{ "type": "code_interpreter" }
162162
],
163163
"model": "gpt-4-1106-preview",
164-
"file_ids": ["assistant-123abc456"]
164+
"tool_resources"{
165+
"code interpreter": {
166+
"file_ids": ["assistant-1234"]
167+
}
168+
}
165169
}'
166170
```
167171

articles/ai-services/translator/language-support.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,7 @@ ms.author: lajanuar
114114
|Mongolian (Traditional)|`mn-Mong`||| | | |
115115
|Myanmar|`my`||| || |
116116
|Nepali|`ne`||| || |
117-
|Norwegian|`nb`||||||
117+
|Norwegian Bokmål|`nb`||||||
118118
|Nyanja|`nya`||| | | |
119119
|Odia|`or`||||| |
120120
|Pashto|`ps`||| || |
@@ -241,7 +241,7 @@ ms.author: lajanuar
241241
|Mongolian (Traditional)|`mn-Mong`|Yes|No|
242242
|Myanmar (Burmese)|`my`|Yes|No|
243243
|Nepali|`ne`|Yes|Yes|
244-
|Norwegian|`nb`|Yes|Yes|
244+
|Norwegian Bokmål|`nb`|Yes|Yes|
245245
|Odia|`or`|Yes|No|
246246
|Pashto|`ps`|Yes|No|
247247
|Persian|`fa`|Yes|No|

articles/machine-learning/how-to-deploy-custom-container.md

Lines changed: 20 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -226,15 +226,30 @@ blue_deployment = ManagedOnlineDeployment(
226226

227227
---
228228

229-
There are a few important concepts to notice in this YAML/Python parameter:
229+
There are a few important concepts to note in this YAML/Python parameter:
230230

231-
#### Readiness route vs. liveness route
231+
#### Base image
232232

233-
An HTTP server defines paths for both _liveness_ and _readiness_. A liveness route is used to check whether the server is running. A readiness route is used to check whether the server is ready to do work. In machine learning inference, a server could respond 200 OK to a liveness request before loading a model. The server could respond 200 OK to a readiness request only after the model is loaded into memory.
233+
The base image is specified as a parameter in environment, and `docker.io/tensorflow/serving:latest` is used in this example. As you inspect the container, you can find that this server uses `ENTRYPOINT` to start an entry point script, which takes the environment variables such as `MODEL_BASE_PATH` and `MODEL_NAME`, and exposes ports such as `8501`. These details are all specific information for this chosen server. You can use this understanding of the server, to determine how to define the deployment. For example, if you set environment variables for `MODEL_BASE_PATH` and `MODEL_NAME` in the deployment definition, the server (in this case, TF Serving) will take the values to initiate the server. Likewise, if you set the port for the routes to be `8501` in the deployment definition, the user request to such routes will be correctly routed to the TF Serving server.
234234

235-
For more information about liveness and readiness probes, see the [Kubernetes documentation](https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/).
235+
Note that this specific example is based on the TF Serving case, but you can use any containers that will stay up and respond to requests coming to liveness, readiness, and scoring routes. You can refer to other examples and see how the dockerfile is formed (for example, using `CMD` instead of `ENTRYPOINT`) to create the containers.
236236

237-
Notice that this deployment uses the same path for both liveness and readiness, since TF Serving only defines a liveness route.
237+
#### Inference config
238+
239+
Inference config is a parameter in environment, and it specifies the port and path for 3 types of the route: liveness, readiness, and scoring route. Inference config is required if you want to run your own container with managed online endpoint.
240+
241+
#### Readiness route vs liveness route
242+
243+
The API server you choose may provide a way to check the status of the server. There are two types of the route that you can specify: _liveness_ and _readiness_. A liveness route is used to check whether the server is running. A readiness route is used to check whether the server is ready to do work. In the context of machine learning inferencing, a server could respond 200 OK to a liveness request before loading a model, and the server could respond 200 OK to a readiness request only after the model is loaded into the memory.
244+
245+
For more information about liveness and readiness probes in general, see the [Kubernetes documentation](https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/).
246+
247+
The liveness and readiness routes will be determined by the API server of your choice, as you would have identified when testing the container locally in earlier step. Note that the example deployment in this article uses the same path for both liveness and readiness, since TF Serving only defines a liveness route. Refer to other examples for different patterns to define the routes.
248+
249+
#### Scoring route
250+
251+
The API server you choose would provide a way to receive the payload to work on. In the context of machine learning inferencing, a server would receive the input data via a specific route. Identify this route for your API server as you test the container locally in earlier step, and specify it when you define the deployment to create.
252+
Note that the successful creation of the deployment will update the scoring_uri parameter of the endpoint as well, which you can verify with `az ml online-endpoint show -n <name> --query scoring_uri`.
238253

239254
#### Locating the mounted model
240255

articles/open-datasets/dataset-1000-genomes.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,8 @@ ms.date: 07/10/2024
99

1010
# 1000 Genomes
1111

12+
[!INCLUDE [Open Dataset access change notice](./includes/open-datasets-change-note.md)]
13+
1214
The 1000 Genomes Project ran between 2008 and 2015, to create the largest public catalog of human variation and genotype data. The final data set contains data for 2,504 individuals from 26 populations and 84 million identified variants. For more information, visit the 1000 Genome Project [website](https://www.internationalgenome.org/) and these publications:
1315

1416
[Pilot Analysis: A map of human genome variation from population-scale sequencing Nature 467, 1061-1073 (28 October 2010)](https://www.nature.com/articles/nature09534)

articles/open-datasets/dataset-clinvar-annotations.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,8 @@ ms.date: 06/13/2024
99

1010
# ClinVar Annotations
1111

12+
[!INCLUDE [Open Dataset access change notice](./includes/open-datasets-change-note.md)]
13+
1214
The [ClinVar](https://www.ncbi.nlm.nih.gov/clinvar/) resource is a freely accessible, public archive of reports - with supporting evidence - about the relationships among human variations and phenotypes. It facilitates access to and communication about the claimed relationships between human variation and observed health status, and about the history of that interpretation. It provides access to a broader set of clinical interpretations that researchers can incorporate into genomics workflows and applications.
1315

1416
Visit the [Data Dictionary](https://www.ncbi.nlm.nih.gov/projects/clinvar/ClinVarDataDictionary.pdf) and the [FAQ resource](https://www.ncbi.nlm.nih.gov/clinvar/docs/faq/) for more information about the data.

articles/open-datasets/dataset-encode.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@ ms.date: 04/16/2021
88

99
# ENCODE: Encyclopedia of DNA Elements
1010

11+
[!INCLUDE [Open Dataset access change notice](./includes/open-datasets-change-note.md)]
12+
1113
The [Encyclopedia of DNA Elements (ENCODE) Consortium](https://www.encodeproject.org/help/project-overview/) is an ongoing international collaboration of research groups funded by the National Human Genome Research Institute (NHGRI). ENCODE's goal is to build a comprehensive parts list of functional elements in the human genome, including elements that act at the protein and RNA levels, and regulatory elements that control cells and circumstances in which a gene is active.
1214

1315
ENCODE investigators employ various assays and methods to identify functional elements. The discovery and annotation of gene elements is accomplished primarily by sequencing a diverse range of RNA sources, comparative genomics, integrative bioinformatic methods, and human curation. Regulatory elements are typically investigated through DNA hypersensitivity assays, assays of DNA methylation, and immunoprecipitation (IP) of proteins that interact with DNA and RNA, that is, modified histones, transcription factors, chromatin regulators, and RNA-binding proteins, followed by sequencing.

articles/open-datasets/dataset-gatk-resource-bundle.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@ ms.date: 04/16/2021
88

99
# GATK Resource Bundle
1010

11+
[!INCLUDE [Open Dataset access change notice](./includes/open-datasets-change-note.md)]
12+
1113
The [GATK resource bundle](https://gatk.broadinstitute.org/hc/articles/360035890811-Resource-bundle) is a collection of standard files for working with human resequencing data with the GATK.
1214

1315
[!INCLUDE [Open Dataset usage notice](./includes/open-datasets-usage-note.md)]

articles/open-datasets/dataset-genomics-data-lake.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,8 @@ The Genomics Data Lake is hosted in the West US 2 and West Central US Azure regi
3030
| [GATK Resource Bundle](dataset-gatk-resource-bundle.md) | GATK Resource bundle |
3131
| [TCGA Open Data](dataset-the-cancer-genome-atlas.md) | TCGA Open Data |
3232
| [Pan UK-Biobank](dataset-panancestry-uk-bio-bank.md) | Pan UK-Biobank |
33+
| [ImmuneCODE database](dataset-immunecode.md) | ImmuneCODE database |
34+
| [Open Targets dataset](dataset-panancestry-uk-bio-bank.md) | Open Targets dataset |
3335

3436
## Next steps
3537

articles/open-datasets/dataset-human-reference-genomes.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@ ms.date: 04/16/2021
88

99
# Human Reference Genomes
1010

11+
[!INCLUDE [Open Dataset access change notice](./includes/open-datasets-change-note.md)]
12+
1113
This dataset includes two human-genome references assembled by the [Genome Reference Consortium](https://www.ncbi.nlm.nih.gov/grc): Hg19 and Hg38.
1214

1315
For more information on Hg19 (GRCh37) data, see the [GRCh37 report at NCBI](https://www.ncbi.nlm.nih.gov/assembly/GCF_000001405.13/).

0 commit comments

Comments
 (0)