Skip to content

Commit 166d097

Browse files
finish (#1501)
1 parent ca7bdec commit 166d097

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

optimize-llm.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -174,7 +174,7 @@ There are various quantization techniques, which we won't discuss in detail here
174174
- 3. Dynamically dequantize weights to bfloat16 to perform the computation with their input vectors in bfloat16 precision
175175
- 4. Quantize the weights again to the target precision after computation with their inputs.
176176

177-
In a nutshell, this means that *inputs-weight matrix* multiplications, with \\( X \\) being the *inputs*, \\( W )\\ being a weight matrix and \\( Y )\\ being the output:
177+
In a nutshell, this means that *inputs-weight matrix* multiplications, with \\( X \\) being the *inputs*, \\( W \\) being a weight matrix and \\( Y \\) being the output:
178178

179179
$$ Y = X * W $$
180180

0 commit comments

Comments
 (0)