Skip to content

Commit e0241ee

Browse files
committed
remove useless
1 parent 4f9000e commit e0241ee

File tree

2 files changed

+3
-9
lines changed

2 files changed

+3
-9
lines changed

test/prototype/inductor/test_int8_sdpa_fusion.py renamed to test/prototype/inductor/test_qsdpa_fusion.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -150,14 +150,14 @@ def _check_common(
150150
@skipIfRocm
151151
@unittest.skipIf(
152152
not torch_version_at_least("2.7.0"),
153-
reason="int8 sdpa requires torch 2.7 or later",
153+
reason="qsdpa requires torch 2.7 or later",
154154
)
155155
@unittest.skipIf(
156156
"CPU" not in torch._C._dispatch_dump("torchao::qscaled_dot_product"),
157157
reason="cpp kernels not built",
158158
)
159159
@config.patch({"freezing": True})
160-
def _test_sdpa_int8_rewriter(self):
160+
def _test_qsdpa_rewriter(self):
161161
import torchao.quantization.pt2e.quantizer.x86_inductor_quantizer as xiq
162162
from torchao.quantization.pt2e.quantize_pt2e import convert_pt2e, prepare_pt2e
163163
from torchao.quantization.pt2e.quantizer.x86_inductor_quantizer import (
@@ -212,9 +212,7 @@ def _test_sdpa_int8_rewriter(self):
212212

213213
class SDPAPatternRewriterCpuTests(TestSDPAPatternRewriterTemplate):
214214
device = "cpu"
215-
test_sdpa_int8_rewriter_cpu = (
216-
TestSDPAPatternRewriterTemplate._test_sdpa_int8_rewriter
217-
)
215+
test_qsdpa_rewriter_cpu = TestSDPAPatternRewriterTemplate._test_qsdpa_rewriter
218216

219217

220218
if __name__ == "__main__":

torchao/csrc/cpu/quantized_sdpa.cpp

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2509,7 +2509,6 @@ at::Tensor _qscaled_dot_product_cpu(
25092509
#ifdef CPU_CAPABILITY_AVX512
25102510
if (at::native::cpublas::could_pack(dtype)) {
25112511
at::Tensor output = at::empty_like(query, query.options()).transpose(1, 2);
2512-
std::cout << "int8_sdpa_fused_kernel" << std::endl;
25132512
int8_sdpa_fused_kernel(output, query, key, value,
25142513
dropout_p, is_causal, attn_mask, scale,
25152514
q_scale, q_zp,
@@ -2520,7 +2519,6 @@ at::Tensor _qscaled_dot_product_cpu(
25202519
return output.transpose(1, 2);
25212520
} else {
25222521
#endif // CPU_CAPABILITY_AVX512
2523-
std::cout << "int8_sdpa_math_kernel" << std::endl;
25242522
return int8_sdpa_math_kernel(query, key, value,
25252523
dropout_p, is_causal, attn_mask, scale,
25262524
q_scale, q_zp,
@@ -2536,7 +2534,6 @@ at::Tensor _qscaled_dot_product_cpu(
25362534
// CPUBLAS_BRGEMM_F8F8F32 is defined if FP8 BRGEMM is supported in PyTorch CPUBlas.
25372535
if (at::native::cpublas::could_pack(dtype)) {
25382536
at::Tensor output = at::empty_like(query, query.options()).transpose(1, 2);
2539-
std::cout << "fp8_sdpa_fused_kernel" << std::endl;
25402537
fp8_sdpa_fused_kernel(output, query, key, value,
25412538
dropout_p, is_causal, attn_mask, scale,
25422539
q_scale, k_scale,
@@ -2545,7 +2542,6 @@ at::Tensor _qscaled_dot_product_cpu(
25452542
return output.transpose(1, 2);
25462543
} else {
25472544
#endif // CPU_CAPABILITY_AVX512 && CPUBLAS_BRGEMM_F8F8F32
2548-
std::cout << "fp8_sdpa_math_kernel" << std::endl;
25492545
return fp8_sdpa_math_kernel(query, key, value,
25502546
dropout_p, is_causal, attn_mask, scale,
25512547
q_scale, k_scale,

0 commit comments

Comments
 (0)