Skip to content

Commit 47abe2c

Browse files
committed
test
1 parent a65f0f1 commit 47abe2c

File tree

1 file changed

+0
-7
lines changed

1 file changed

+0
-7
lines changed

tools/llm/torchtrt_ext/sdpa_converter.py

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -201,13 +201,6 @@ def scaled_dot_product_attention(
201201
temp_mask = impl.unary.logical_not(
202202
ctx, target, source_ir, name + "_logical_not", tril_tensor
203203
)
204-
temp_mask = cast_trt_tensor(
205-
ctx, temp_mask, trt.float32, name + "_casted_bool", target, source_ir
206-
)
207-
temp_mask = impl.elementwise.mul(
208-
ctx, target, source_ir, name + "_mul_-inf", temp_mask, float("-inf")
209-
)
210-
attn_bias = temp_mask
211204

212205
# This need_mask determines if we want to use the causal mask or not
213206
# When KV caching is enabled, L = 1 and != S. In this case, we shouldn't use the causal mask.

0 commit comments

Comments
 (0)