Skip to content

Commit fd92d96

Browse files
committed
New conceptual doc on language models for AI and machine learning
1 parent 900c630 commit fd92d96

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

articles/aks/concepts-ai-ml-language-models.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: Concepts - Small and large language models
33
description: Learn about small and large language models, including when and how you can use them with your Azure Kubernetes Service (AKS) AI and machine learning workloads.
44
ms.topic: conceptual
5-
ms.date: 06/14/2024
5+
ms.date: 06/17/2024
66
author: schaffererin
77
ms.author: schaffererin
88
---
@@ -16,11 +16,11 @@ In this article, you learn about small and large language models, including when
1616

1717
Language models are powerful machine learning models used for natural language processing (NLP) tasks, such as text generation and sentiment analysis. These models represent natural language based on the probability of words or sequences of words occurring in a given context.
1818

19-
*Conventional language models* have been used in supervised settings where they're trained on well-labeled text datasets for specific tasks. *Pretrained language models*, on the other hand, are trained on large-scale text corpora from the internet using deep neural networks and can be fine-tuned on smaller datasets for specific tasks.
19+
*Conventional language models* are used in supervised settings where they're trained on well-labeled text datasets for specific tasks. *Pretrained language models* are trained on large-scale text corpora from the internet using deep neural networks and can be fine-tuned on smaller datasets for specific tasks.
2020

2121
The size of a language model is determined by the number of parameters it has, which is the number of weights that determine how the model processes input data and generates output. Parameters are learned during the training process by adjusting the weights to minimize the difference between the model's predictions and the actual data. The more parameters a model has, the more complex and expressive it is, but also the more computationally expensive it is to train and use.
2222

23-
In general, ***small language models*** have *fewer than 100 million parameters*, while ***large language models*** have *more than 100 million parameters*. For example, GPT-2 has four versions with different sizes: small (124 million parameters), medium (355 million parameters), large (774 million parameters), and extra-large (1.5 billion parameters).
23+
In general, **small language models** have *fewer than 100 million parameters*, and **large language models** have *more than 100 million parameters*. For example, GPT-2 has four versions with different sizes: small (124 million parameters), medium (355 million parameters), large (774 million parameters), and extra-large (1.5 billion parameters).
2424

2525
## When to use small language models
2626

0 commit comments

Comments
 (0)