Skip to content

Commit 85c0ad7

Browse files
committed
remove useless
1 parent 36b9775 commit 85c0ad7

File tree

2 files changed

+3
-9
lines changed

2 files changed

+3
-9
lines changed

test/prototype/inductor/test_int8_sdpa_fusion.py renamed to test/prototype/inductor/test_qsdpa_fusion.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -151,14 +151,14 @@ def _check_common(
151151
@skipIfRocm
152152
@unittest.skipIf(
153153
not torch_version_at_least("2.7.0"),
154-
reason="int8 sdpa requires torch 2.7 or later",
154+
reason="qsdpa requires torch 2.7 or later",
155155
)
156156
@unittest.skipIf(
157157
"CPU" not in torch._C._dispatch_dump("torchao::qscaled_dot_product"),
158158
reason="cpp kernels not built",
159159
)
160160
@config.patch({"freezing": True})
161-
def _test_sdpa_int8_rewriter(self):
161+
def _test_qsdpa_rewriter(self):
162162
import torchao.quantization.pt2e.quantizer.x86_inductor_quantizer as xiq
163163
from torchao.quantization.pt2e.quantize_pt2e import convert_pt2e, prepare_pt2e
164164
from torchao.quantization.pt2e.quantizer.x86_inductor_quantizer import (
@@ -213,9 +213,7 @@ def _test_sdpa_int8_rewriter(self):
213213

214214
class SDPAPatternRewriterCpuTests(TestSDPAPatternRewriterTemplate):
215215
device = "cpu"
216-
test_sdpa_int8_rewriter_cpu = (
217-
TestSDPAPatternRewriterTemplate._test_sdpa_int8_rewriter
218-
)
216+
test_qsdpa_rewriter_cpu = TestSDPAPatternRewriterTemplate._test_qsdpa_rewriter
219217

220218

221219
if __name__ == "__main__":

torchao/csrc/cpu/quantized_sdpa.cpp

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2509,7 +2509,6 @@ at::Tensor _qscaled_dot_product_cpu(
25092509
#ifdef CPU_CAPABILITY_AVX512
25102510
if (at::native::cpublas::could_pack(dtype)) {
25112511
at::Tensor output = at::empty_like(query, query.options()).transpose(1, 2);
2512-
std::cout << "int8_sdpa_fused_kernel" << std::endl;
25132512
int8_sdpa_fused_kernel(output, query, key, value,
25142513
dropout_p, is_causal, attn_mask, scale,
25152514
q_scale, q_zp,
@@ -2520,7 +2519,6 @@ at::Tensor _qscaled_dot_product_cpu(
25202519
return output.transpose(1, 2);
25212520
} else {
25222521
#endif // CPU_CAPABILITY_AVX512
2523-
std::cout << "int8_sdpa_math_kernel" << std::endl;
25242522
return int8_sdpa_math_kernel(query, key, value,
25252523
dropout_p, is_causal, attn_mask, scale,
25262524
q_scale, q_zp,
@@ -2536,7 +2534,6 @@ at::Tensor _qscaled_dot_product_cpu(
25362534
// CPUBLAS_BRGEMM_F8F8F32 is defined if FP8 BRGEMM is supported in PyTorch CPUBlas.
25372535
if (at::native::cpublas::could_pack(dtype)) {
25382536
at::Tensor output = at::empty_like(query, query.options()).transpose(1, 2);
2539-
std::cout << "fp8_sdpa_fused_kernel" << std::endl;
25402537
fp8_sdpa_fused_kernel(output, query, key, value,
25412538
dropout_p, is_causal, attn_mask, scale,
25422539
q_scale, k_scale,
@@ -2545,7 +2542,6 @@ at::Tensor _qscaled_dot_product_cpu(
25452542
return output.transpose(1, 2);
25462543
} else {
25472544
#endif // CPU_CAPABILITY_AVX512 && CPUBLAS_BRGEMM_F8F8F32
2548-
std::cout << "fp8_sdpa_math_kernel" << std::endl;
25492545
return fp8_sdpa_math_kernel(query, key, value,
25502546
dropout_p, is_causal, attn_mask, scale,
25512547
q_scale, k_scale,

0 commit comments

Comments
 (0)