Skip to content

Commit e3d3784

Browse files
committed
update
1 parent 15df2fe commit e3d3784

File tree

2 files changed

+15
-13
lines changed

2 files changed

+15
-13
lines changed

articles/ai-services/openai/concepts/customizing-llm.md

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -15,25 +15,25 @@ recommendations: false
1515

1616
There are several techniques for adapting a pre-trained language model to suit a specific task or domain. These include prompt engineering, RAG (Retrieval Augmented Generation), and fine-tuning. These three techniques are not mutually exclusive but are complementary methods that in combination can be applicable to a specific use case. In this article, we'll explore these techniques, illustrative use cases, things to consider, and provide links to resources to learn more and get started with each.
1717

18-
## Techniques
18+
## Prompt engineering
1919

20-
### Definition - prompt engineering
20+
### Definition
2121

2222
[Prompt engineering](./prompt-engineering.md) is a technique that is both art and science, which involves designing prompts for generative AI models. This process utilizes in-context learning ([zero shot and few shot](./prompt-engineering.md#examples)) and, with iteration, improves accuracy and relevancy in responses, optimizing the performance of the model.
2323

24-
### Illustrative use cases - prompt engineering
24+
### Illustrative use cases
2525

2626
A Marketing Manager at an environmentally conscious company can use prompt engineering to help guide the model to generate descriptions that are more aligned with their brand’s tone and style. For instance, they can add a prompt like "Write a product description for a new line of eco-friendly cleaning products that emphasizes quality, effectiveness, and highlights the use of environmentally friendly ingredients" to the input. This will help the model generate descriptions that are aligned with their brand’s values and messaging.
2727

28-
### Things to consider - prompt engineering
28+
### Things to consider
2929

3030
- **Prompt engineering** is the starting point for generating desired output from generative AI models.
3131

3232
- **Craft clear instructions**: Instructions are commonly used in prompts and guide the model's behavior. Be specific and leave as little room for interpretation as possible. Use analogies and descriptive language to help the model understand your desired outcome.
3333

3434
- **Experiment and iterate**: Prompt engineering is an art that requires experimentation and iteration. Practice and gain experience in crafting prompts for different tasks. Every model might behave differently, so it's important to adapt prompt engineering techniques accordingly.
3535

36-
### Getting started - prompt engineering
36+
### Getting started
3737

3838
- [Introduction to prompt engineering](./prompt-engineering.md)
3939
- [Prompt engineering techniques](./advanced-prompt-engineering.md)
@@ -42,45 +42,45 @@ A Marketing Manager at an environmentally conscious company can use prompt engin
4242

4343
## RAG (Retrieval Augmented Generation)
4444

45-
### Definition - RAG
45+
### Definition
4646

4747
[RAG (Retrieval Augmented Generation)](../../../ai-studio/concepts/retrieval-augmented-generation.md) is a method that integrates external data into a Large Language Model prompt to generate relevant responses. This approach is particularly beneficial when using a large corpus of unstructured text based on different topics. It allows for answers to be grounded in the organization’s knowledge base (KB), providing a more tailored and accurate response.
4848

4949
RAG is also advantageous when answering questions based on an organization’s private data or when the public data that the model was trained on might have become outdated. This helps ensure that the responses are always up-to-date and relevant, regardless of the changes in the data landscape.
5050

51-
### Illustrative use case - RAG
51+
### Illustrative use case
5252

5353
A corporate HR department is looking to provide an intelligent assistant that answers specific employee health insurance related questions such as "are eyeglasses covered?" RAG is used to ingest the extensive and numerous documents associated with insurance plan policies to enable the answering of these specific types of questions.
5454

55-
### Things to consider - RAG
55+
### Things to consider
5656

5757
- RAG helps ground AI output in real-world data and reduces the likelihood of fabrication.
5858

5959
- RAG is helpful when there is a need to answer questions based on private proprietary data.
6060

6161
- RAG is helpful when you might want questions answered that are recent (for example, before the cutoff date of when the [model version](./models.md) was last trained).
6262

63-
### Getting started - RAG
63+
### Getting started
6464

6565
- [Retrieval Augmented Generation in Azure AI Studio - Azure AI Studio | Microsoft Learn](../../../ai-studio/concepts/retrieval-augmented-generation.md)
6666
- [Retrieval Augmented Generation (RAG) in Azure AI Search](../../../search/retrieval-augmented-generation-overview.md)
6767
- [Retrieval Augmented Generation using Azure Machine Learning prompt flow (preview)](../../../machine-learning/concept-retrieval-augmented-generation.md)
6868

6969
## Fine-tuning
7070

71-
### Definition - fine-tuning
71+
### Definition
7272

7373
[Fine-tuning](../how-to/fine-tuning.md), specifically [supervised fine-tuning](https://techcommunity.microsoft.com/t5/ai-azure-ai-services-blog/fine-tuning-now-available-with-azure-openai-service/ba-p/3954693?lightbox-message-images-3954693=516596iC5D02C785903595A) in this context, is an iterative process that adapts an existing large language model to a provided training set in order to improve performance, teach the model new skills, or reduce latency. This approach is used when the model needs to learn and generalize over specific topics, particularly when these topics are generally small in scope.
7474

7575
Fine-tuning requires the use of high-quality training data, in a [special example based format](../how-to/fine-tuning.md#example-file-format), to create the new fine-tuned Large Language Model. By focusing on specific topics, fine-tuning allows the model to provide more accurate and relevant responses within those areas of focus.
7676

77-
### Illustrative use case - fine-tuning
77+
### Illustrative use case
7878

7979
An IT department has been using GPT-4 to convert natural language queries to SQL, but they have found that the responses are not always reliably grounded in their schema, and the cost is prohibitively high.
8080

8181
They fine-tune GPT-3.5-Turbo with hundreds of requests and correct responses and produce a model that performs better than the base model with lower costs and latency.
8282

83-
### Things to consider - fine-tuning
83+
### Things to consider
8484

8585
- Fine-tuning is an advanced capability; it enhances LLM with after-cutoff-date knowledge and/or domain specific knowledge. Start by evaluating the baseline performance of a standard model against their requirements before considering this option.
8686

@@ -94,7 +94,7 @@ They fine-tune GPT-3.5-Turbo with hundreds of requests and correct responses and
9494

9595
- Fine-tuning has upfront costs for training the model. And additional hourly costs for hosting the custom model once it's deployed.
9696

97-
## Getting started - fine-tuning
97+
## Getting started
9898

9999
- [When to use Azure OpenAI fine-tuning](./fine-tuning-considerations.md)
100100
- [Customize a model with fine-tuning](../how-to/fine-tuning.md)

articles/ai-services/openai/toc.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -47,6 +47,8 @@ items:
4747
href: ./concepts/abuse-monitoring.md
4848
- name: Content filtering
4949
href: ./concepts/content-filter.md
50+
- name: Customizing Large Language Models (LLMs)
51+
href: ./concepts/customizing-llms.md
5052
- name: Embeddings
5153
href: ./concepts/understand-embeddings.md
5254
- name: Fine-tuning

0 commit comments

Comments
 (0)