We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent fe8ee0c commit fb2e693Copy full SHA for fb2e693
python/triton_kernels/triton_kernels/matmul_ogs_details/_matmul_ogs.py
@@ -301,7 +301,6 @@ def _matmul_ogs(
301
w_scale = load_scale(WScale)
302
acc *= x_scale * w_scale
303
acc = acc + bias[None, :] * betas[:, None]
304
- acc *= gammas[:, None]
305
if out_alpha is not None:
306
acc *= out_alpha
307
if ACTIVATION_FN is not None:
@@ -312,6 +311,7 @@ def _matmul_ogs(
312
311
else:
313
tl.static_assert(ACTIVATION_REDUCTION_N == 1, "Activation reduction must be 1 if no activation fn is provided")
314
out = acc
+ out *= gammas[:, None]
315
# write-back
316
Y += start_z.to(index_type) * stride_y_z
317
if WriteBackIndx is not None:
0 commit comments