Skip to content

Commit d6d38d3

Browse files
Merge pull request #49336 from GraemeMalcolm/main
Updated embeddings info for accuracy
2 parents 00a8a91 + b0c96dc commit d6d38d3

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

learn-pr/wwl-data-ai/analyze-images-computer-vision/includes/2b-computer-vision-models.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -32,13 +32,13 @@ CNNs have been at the core of computer vision solutions for many years. While th
3232

3333
### Transformers
3434

35-
Most advances in computer vision over the decades have been driven by improvements in CNN-based models. However, in another AI discipline - *natural language processing* (NLP), another type of neural network architecture, called a *transformer* has enabled the development of sophisticated models for language. Transformers work by processing huge volumes of data, and encoding language *tokens* (representing individual words or phrases) as vector-based *embeddings* (arrays of numeric values). You can think of an embedding as representing a set of dimensions that each represent some semantic attribute of the token. The embeddings are created such that tokens that are commonly used in the same context are closer together dimensionally than unrelated words.
35+
Most advances in computer vision over the decades have been driven by improvements in CNN-based models. However, in another AI discipline - *natural language processing* (NLP), another type of neural network architecture, called a *transformer* has enabled the development of sophisticated models for language. Transformers work by processing huge volumes of data, and encoding language *tokens* (representing individual words or phrases) as vector-based *embeddings* (arrays of numeric values). You can think of an embedding as representing a set of dimensions that each represent some semantic attribute of the token. The embeddings are created such that tokens that are commonly used in the same context define vectors that are more closely aligned than unrelated words.
3636

3737
As a simple example, the following diagram shows some words encoded as three-dimensional vectors, and plotted in a 3D space:
3838

3939
![Diagram of token vectors in a 3D space.](../media/language-encoder.png)
4040

41-
Tokens that are semantically similar are encoded in similar positions, creating a semantic language model that makes it possible to build sophisticated NLP solutions for text analysis, translation, language generation, and other tasks.
41+
Tokens that are semantically similar are encoded in similar directions, creating a semantic language model that makes it possible to build sophisticated NLP solutions for text analysis, translation, language generation, and other tasks.
4242

4343
> [!NOTE]
4444
> We've used only three dimensions, because that's easy to visualize. In reality, encoders in transformer networks create vectors with many more dimensions, defining complex semantic relationships between tokens based on linear algebraic calculations. The math involved is complex, as is the architecture of a transformer model. Our goal here is just to provide a *conceptual* understanding of how encoding creates a model that encapsulates relationships between entities.
4.05 KB
Loading

0 commit comments

Comments
 (0)