We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 838f08c commit 444e0a5Copy full SHA for 444e0a5
posts/13-10-25-klevel-opt/index.html
@@ -28,7 +28,7 @@ <h1>L2L: Layer-wise 2-Level Optimization</h1>
28
<h3>Background</h3>
29
30
<p>
31
- This post outlines a training procedure for large language models (LLMs)—and neural networks in general—that I
+ This post outlines a training procedure for large language models (LLMs)---and neural networks in general—that I
32
call <strong>Layer-wise 2-Level Optimization (L2L)</strong>. The idea is inspired by the
33
<em>K-Level Policy Gradient</em> paper <span class="cite" value="Reddi25"></span>, which is well worth a read if
34
you're into multi-agent learning.
0 commit comments