Skip to content

Commit 455b6c6

Browse files
authored
Update 6.mdx
Fixes two typos
1 parent e2698dc commit 455b6c6

File tree

1 file changed

+3
-3
lines changed
  • chapters/en/chapter1

1 file changed

+3
-3
lines changed

chapters/en/chapter1/6.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55

66
# Transformer Architectures[[transformer-architectures]]
77

8-
In the previous sections, we introduced the general Transformer architecture and explored how these models can solve various tasks. Now, let's take a closer look at the three main architectural variants of Transformer models and understand when to use each one. Then, we looked at how those architectures are applied to different language tasks.
8+
In the previous sections, we introduced the general Transformer architecture and explored how these models can solve various tasks. Now, let's take a closer look at the three main architectural variants of Transformer models and understand when to use each one. Then, we look at how those architectures are applied to different language tasks.
99

1010
In this section, we're going to dive deeper into the three main architectural variants of Transformer models and understand when to use each one.
1111

@@ -85,7 +85,7 @@ Modern decoder-based LLMs have demonstrated impressive capabilities:
8585
| Reasoning | Working through problems step by step | Solving math problems or logical puzzles |
8686
| Few-shot learning | Learning from a few examples in the prompt | Classifying text after seeing just 2-3 examples |
8787

88-
You can experiment with decoder-based LLMs directly in your browser via model repo pages on the Hub. Here's an an example with the classic [GPT-2](https://huggingface.co/openai-community/gpt2) (OpenAI's finest open source model!):
88+
You can experiment with decoder-based LLMs directly in your browser via model repo pages on the Hub. Here's an example with the classic [GPT-2](https://huggingface.co/openai-community/gpt2) (OpenAI's finest open source model!):
8989

9090
<iframe
9191
src="https://huggingface.co/openai-community/gpt2"
@@ -221,4 +221,4 @@ in E2.
221221

222222
In this section, we've explored the three main Transformer architectures and some specialized attention mechanisms. Understanding these architectural differences is crucial for selecting the right model for your specific NLP task.
223223

224-
As we move forward in the course, you'll get hands-on experience with these different architectures and learn how to fine-tune them for your specific needs. In the next section, we'll look at some of the limitations and biases present in these models that you should be aware of when deploying them.
224+
As we move forward in the course, you'll get hands-on experience with these different architectures and learn how to fine-tune them for your specific needs. In the next section, we'll look at some of the limitations and biases present in these models that you should be aware of when deploying them.

0 commit comments

Comments
 (0)