Skip to content

Commit bcb4d86

Browse files
authored
Merge pull request #274410 from shubhirajMsft/patch-9
Update and rename model-catalog.md to model-catalog-overview.md
2 parents 55496da + c895ba1 commit bcb4d86

23 files changed

+247
-121
lines changed

.openpublishing.redirection.json

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3430,6 +3430,12 @@
34303430
"redirect_url": "/azure/ai-services/anomaly-detector/quickstarts/detect-data-anomalies-python",
34313431
"redirect_document_id": false
34323432
},
3433+
{
3434+
"source_path_from_root": "/articles/ai-studio/how-to/model-catalog.md",
3435+
"redirect_url": "/azure/ai-studio/how-to/model-catalog-overview",
3436+
"redirect_document_id": false
3437+
},
3438+
34333439
{
34343440
"source_path_from_root": "/articles/service-fabric/upgrade-managed-disks.md",
34353441
"redirect_url": "/azure/service-fabric/service-fabric-scale-up-primary-node-type",

articles/ai-studio/concepts/ai-resources.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ With the same API key, you can access all of the following Azure AI services:
8484
| ![Speech icon](../../ai-services/media/service-icons/speech.svg) [Speech](../../ai-services/speech-service/index.yml) | Speech to text, text to speech, translation and speaker recognition |
8585
| ![Vision icon](../../ai-services/media/service-icons/vision.svg) [Vision](../../ai-services/computer-vision/index.yml) | Analyze content in images and videos |
8686

87-
Large language models that can be used to generate text, speech, images, and more, are hosted by the Azure AI hub resource. Fine-tuned models and open models deployed from the [model catalog](../how-to/model-catalog.md) are always created in the project context for isolation.
87+
Large language models that can be used to generate text, speech, images, and more, are hosted by the Azure AI hub resource. Fine-tuned models and open models deployed from the [model catalog](../how-to/model-catalog-overview.md) are always created in the project context for isolation.
8888

8989
### Virtual networking
9090

articles/ai-studio/concepts/content-filtering.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ This system is powered by [Azure AI Content Safety](../../ai-services/content-sa
2626

2727
The content filtering models have been trained and tested on the following languages: English, German, Japanese, Spanish, French, Italian, Portuguese, and Chinese. However, the service can work in many other languages, but the quality can vary. In all cases, you should do your own testing to ensure that it works for your application.
2828

29-
You can create a content filter or use the default content filter for Azure OpenAI model deployment, and can also use a default content filter for other text models curated by Azure AI in the [model catalog](../how-to/model-catalog.md). The custom content filters for those models aren't yet available. Models available through Models as a Service have content filtering enabled by default and can't be configured.
29+
You can create a content filter or use the default content filter for Azure OpenAI model deployment, and can also use a default content filter for other text models curated by Azure AI in the [model catalog](../how-to/model-catalog-overview.md). The custom content filters for those models aren't yet available. Models available through Models as a Service have content filtering enabled by default and can't be configured.
3030

3131
## How to create a content filter?
3232
For any model deployment in [Azure AI Studio](https://ai.azure.com), you could directly use the default content filter, but when you want to have more customized setting on content filter, for example set a stricter or looser filter, or enable more advanced capabilities, like jailbreak risk detection and protected material detection. To create a content filter, you could go to **Build**, choose one of your projects, then select **Content filters** in the left navigation bar, and create a content filter.
@@ -44,9 +44,9 @@ The content filtering system integrated in Azure AI Studio contains neural multi
4444
|Category|Description|
4545
|--------|-----------|
4646
| Hate |The hate category describes language attacks or uses that include pejorative or discriminatory language with reference to a person or identity group based on certain differentiating attributes of these groups including but not limited to race, ethnicity, nationality, gender identity and expression, sexual orientation, religion, immigration status, ability status, personal appearance, and body size. |
47-
| Sexual | The sexual category describes language related to anatomical organs and genitals, romantic relationships, acts portrayed in erotic or affectionate terms, physical sexual acts, including those portrayed as an assault or a forced sexual violent act against ones will, prostitution, pornography, and abuse. |
47+
| Sexual | The sexual category describes language related to anatomical organs and genitals, romantic relationships, acts portrayed in erotic or affectionate terms, physical sexual acts, including those portrayed as an assault or a forced sexual violent act against one's will, prostitution, pornography, and abuse. |
4848
| Violence | The violence category describes language related to physical actions intended to hurt, injure, damage, or kill someone or something; describes weapons, etc. |
49-
| Self-Harm | The self-harm category describes language related to physical actions intended to purposely hurt, injure, or damage ones body, or kill oneself.|
49+
| Self-Harm | The self-harm category describes language related to physical actions intended to purposely hurt, injure, or damage one's body, or kill oneself.|
5050

5151
#### Severity levels
5252

articles/ai-studio/concepts/deployments-overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ You often hear this interaction with a model referred to as "inferencing". Infer
2525

2626
First you might ask:
2727
- "What models can I deploy?" Azure AI Studio supports deploying some of the most popular large language and vision foundation models curated by Microsoft, Hugging Face, and Meta.
28-
- "How do I choose the right model?" Azure AI Studio provides a [model catalog](../how-to/model-catalog.md) that allows you to search and filter models based on your use case. You can also test a model on a sample playground before deploying it to your project.
28+
- "How do I choose the right model?" Azure AI Studio provides a [model catalog](../how-to/model-catalog-overview.md) that allows you to search and filter models based on your use case. You can also test a model on a sample playground before deploying it to your project.
2929
- "From where in Azure AI Studio can I deploy a model?" You can deploy a model from the model catalog or from your project's deployment page.
3030

3131
Azure AI Studio simplifies deployments. A simple select or a line of code deploys a model and generate an API endpoint for your applications to consume.

articles/ai-studio/concepts/evaluation-improvement-strategies.md

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -23,17 +23,17 @@ Mitigating content risks and poor quality generations presented by large languag
2323

2424
## Model layer
2525

26-
At the model level, it's important to understand the models you'll be use and what fine-tuning steps might have been taken by the model developers to align the model towards its intended uses and to reduce the risk of potentially risky uses and outcomes. For example, we have collaborated with OpenAI on using techniques such as Reinforcement learning from human feedback (RLHF) and fine-tuning in the base models to build safety into the model itself, and you see safety built into the model to mitigate unwanted behaviors.
26+
At the model level, it's important to understand the models you'll use and what fine-tuning steps might have been taken by the model developers to align the model towards its intended uses and to reduce the risk of potentially risky uses and outcomes. For example, we have collaborated with OpenAI on using techniques such as Reinforcement learning from human feedback (RLHF) and fine-tuning in the base models to build safety into the model itself, and you see safety built into the model to mitigate unwanted behaviors.
2727

28-
Besides these enhancements, Azure AI Studio also offers model catalog that enables you to better understand each model’s capabilities before you even start building your AI applications. You can explore models from Azure OpenAI Service, Meta, etc., organized by collection and task. In the [model catalog](../how-to/model-catalog.md), you can explore model cards to understand model capabilities and limitations, and any safety fine-tuning performed. You can further run sample inferences to see how a model’s responds to typical prompts for a specific use case and experiment with sample inferences.
28+
Besides these enhancements, Azure AI Studio also offers a model catalog that enables you to better understand the capabilities of each model before you even start building your AI applications. You can explore models from Azure OpenAI Service, Meta, etc., organized by collection and task. In the [model catalog](../how-to/model-catalog-overview.md), you can explore model cards to understand model capabilities and limitations and any safety fine-tuning performed. You can further run sample inferences to see how a model responds to typical prompts for a specific use case and experiment with sample inferences.
2929

30-
The model catalog also provides model benchmarks to help users compare each models accuracy using public datasets.
30+
The model catalog also provides model benchmarks to help users compare each model's accuracy using public datasets.
3131

3232
The catalog has over 1,600 models today, including leading models from OpenAI, Mistral, Meta, Hugging Face, and Microsoft.
3333

3434
## Safety systems layer
3535

36-
Choosing a great base model is just the first step. For most AI applications, its not enough to rely on the safety mitigations built into the model itself. Even with fine-tuning, LLMs can make mistakes and are susceptible to attacks such as jailbreaks. In many applications at Microsoft, we use another AI-based safety system, [Azure AI Content Safety](https://azure.microsoft.com/products/ai-services/ai-content-safety/), to provide an independent layer of protection, helping you to block the output of risky content. Azure AI Content Safety is a content moderation offering that goes around the model and monitors the inputs and outputs to help identify and prevent attacks from being successful and catches places where the models make a mistake.
36+
Choosing a great base model is just the first step. For most AI applications, it's not enough to rely on the safety mitigations built into the model itself. Even with fine-tuning, LLMs can make mistakes and are susceptible to attacks such as jailbreaks. In many applications at Microsoft, we use another AI-based safety system, [Azure AI Content Safety](https://azure.microsoft.com/products/ai-services/ai-content-safety/), to provide an independent layer of protection, helping you to block the output of risky content. Azure AI Content Safety is a content moderation offering that goes around the model and monitors the inputs and outputs to help identify and prevent attacks from being successful and catches places where the models make a mistake.
3737

3838
When you deploy your model through the model catalog or deploy your LLM applications to an endpoint, you can use [Azure AI Content Safety](../concepts/content-filtering.md). This safety system works by running both the prompt and completion for your model through an ensemble of classification models aimed at detecting and preventing the output of harmful content across a range of [categories](/azure/ai-services/content-safety/concepts/harm-categories):
3939

@@ -46,31 +46,31 @@ The default configuration is set to filter risky content at the medium severity
4646

4747
## Metaprompt and grounding layer
4848

49-
System message (otherwise known as metaprompt) design and proper data grounding are at the heart of every generative AI application. They provide an applications unique differentiation and are also a key component in reducing errors and mitigating risks. At Microsoft, we find [retrieval augmented generation](./retrieval-augmented-generation.md) (RAG) to be an effective and flexible architecture. With RAG, you enable your application to retrieve relevant knowledge from selected data and incorporate it into your system message to the model. In this pattern, rather than using the model to store information, which can change over time and based on context, the model functions as a reasoning engine over the data provided to it during the query. This improves the freshness, accuracy, and relevancy of inputs and outputs. In other words, RAG can ground your model in relevant data for more relevant results.
49+
System message (otherwise known as metaprompt) design and proper data grounding are at the heart of every generative AI application. They provide an application's unique differentiation and are also a key component in reducing errors and mitigating risks. At Microsoft, we find [retrieval augmented generation](./retrieval-augmented-generation.md) (RAG) to be an effective and flexible architecture. With RAG, you enable your application to retrieve relevant knowledge from selected data and incorporate it into your system message to the model. In this pattern, rather than using the model to store information, which can change over time and based on context, the model functions as a reasoning engine over the data provided to it during the query. This improves the freshness, accuracy, and relevancy of inputs and outputs. In other words, RAG can ground your model in relevant data for more relevant results.
5050

51-
Now the other part of the story is how you teach the base model to use that data or to answer the questions effectively in your application. When you create a system message, youre giving instructions to the model in natural language to consistently guide its behavior on the backend. Tapping into the trained data of the models is valuable but enhancing it with your information is critical.
51+
Now the other part of the story is how you teach the base model to use that data or to answer the questions effectively in your application. When you create a system message, you're giving instructions to the model in natural language to consistently guide its behavior on the backend. Tapping into the trained data of the models is valuable but enhancing it with your information is critical.
5252

53-
Heres what a system message should look like. You must:
53+
Here's what a system message should look like. You must:
5454

55-
- Define the models profile, capabilities, and limitations for your scenario.
56-
- Define the models output format.
55+
- Define the model's profile, capabilities, and limitations for your scenario.
56+
- Define the model's output format.
5757
- Provide examples to demonstrate the intended behavior of the model.
5858
- Provide additional behavioral guardrails.
5959

6060
Recommended System Message Framework:
6161

62-
- Define the models profile, capabilities, and limitations for your scenario.
63-
- **Define the specific task(s)** you would like the model to complete. Describe who the end users will be, what inputs will be provided to the model, and what you expect the model to output.
64-
- **Define how the model should complete the task**, including any additional tools (like APIs, code, plug-ins) the model can use.
62+
- Define the model's profile, capabilities, and limitations for your scenario.
63+
- **Define the specific task(s)** you would like the model to complete. Describe who the end users are, what inputs are provided to the model, and what you expect the model to output.
64+
- **Define how the model should complete the task**, including any extra tools (like APIs, code, plug-ins) the model can use.
6565
- **Define the scope and limitations** of the model's performance by providing clear instructions.
6666
- **Define the posture and tone** the model should exhibit in its responses.
67-
- Define the models output format.
67+
- Define the model's output format.
6868
- **Define the language and syntax** of the output format. For example, if you want the output to be machine parse-able, you may want tot structure the output to be in JSON, XSON orXML.
6969
- **Define any styling or formatting** preferences for better user readability like bulleting or bolding certain parts of the response
7070
- Provide examples to demonstrate the intended behavior of the model
71-
- **Describe difficult use cases** where the prompt is ambiguous or complicated, to give the model additional visibility into how to approach such cases.
71+
- **Describe difficult use cases** where the prompt is ambiguous or complicated, to give the model more visibility into how to approach such cases.
7272
- **Show chain-of-thought** reasoning to better inform the model on the steps it should take to achieve the desired outcomes.
73-
- Provide additional behavioral guardrails
73+
- Provide more behavioral guardrails
7474
- **Define specific behaviors and safety mitigations** to mitigate risks that have been identified and prioritized for the scenario.
7575

7676
Here we outline a set of best practices instructions you can use to augment your task-based system message instructions to minimize different content risks:
@@ -91,7 +91,7 @@ Here we outline a set of best practices instructions you can use to augment your
9191
### Sample system message instructions for ungrounded answers
9292

9393
```
94-
- Your answer **must not** include any speculation or inference about the background of the document or the users gender, ancestry, roles, positions, etc.
94+
- Your answer **must not** include any speculation or inference about the background of the document or the user's gender, ancestry, roles, positions, etc.
9595
- You **must not** assume or change dates and times.
9696
- You **must always** perform searches on [insert relevant documents that your feature can search on] when the user is seeking information (explicitly or implicitly), regardless of internal knowledge or information.
9797
```

0 commit comments

Comments
 (0)