@@ -858,6 +858,8 @@ module attributes {"triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 1 :
858858module attributes {" triton_gpu.num-ctas" = 1 : i32 , " triton_gpu.num-warps" = 4 : i32 } {
859859 // CHECK-LABEL: convert_layout_slice_mmav2_blocked_reg
860860 tt.func @convert_layout_slice_mmav2_blocked_reg (%arg0: tensor <1 xf16 , #slice >) {
861+ // CHECK-NOT: st.shared
862+ // CHECK-NOT: llvm.load
861863 %0 = triton_gpu.convert_layout %arg0 : tensor <1 xf16 , #slice > -> tensor <1 xf16 , #blocked >
862864 tt.return
863865 }
@@ -901,6 +903,8 @@ module attributes {"triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 4 :
901903module attributes {" triton_gpu.num-ctas" = 1 : i32 , " triton_gpu.num-warps" = 4 : i32 } {
902904 // CHECK-LABEL: convert_layout_mmav3_mmav3_2
903905 tt.func @convert_layout_mmav3_mmav3_2 (%arg0: tensor <16 x16 xf16 , #mma1 >) {
906+ // CHECK-NOT: st.shared
907+ // CHECK-NOT: llvm.load
904908 %0 = triton_gpu.convert_layout %arg0 : tensor <16 x16 xf16 , #mma1 > -> tensor <16 x16 xf16 , #mma0 >
905909 tt.return
906910 }
@@ -915,6 +919,7 @@ module attributes {"triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 4 :
915919 // CHECK-LABEL: convert_layout_mmav3_mmav3_3
916920 tt.func @convert_layout_mmav3_mmav3_3 (%arg0: tensor <1 x64 xf16 , #mma1 >) {
917921 // CHECK-NOT: st.shared
922+ // CHECK-NOT: llvm.load
918923 %0 = triton_gpu.convert_layout %arg0 : tensor <1 x64 xf16 , #mma1 > -> tensor <1 x64 xf16 , #mma0 >
919924 tt.return
920925 }
0 commit comments