@@ -458,20 +458,20 @@ define <8 x i32> @test_v8i32(<8 x i32>* %V) {
458
458
;
459
459
; AVXONLY-LABEL: test_v8i32:
460
460
; AVXONLY: # BB#0: # %entry
461
- ; AVXONLY-NEXT: vmovdqu (%rdi), %ymm0
461
+ ; AVXONLY-NEXT: vmovdqa (%rdi), %ymm0
462
462
; AVXONLY-NEXT: retq
463
463
;
464
464
; KNL-LABEL: test_v8i32:
465
465
; KNL: # BB#0: # %entry
466
- ; KNL-NEXT: vmovdqu (%rdi), %ymm0
466
+ ; KNL-NEXT: vmovdqa (%rdi), %ymm0
467
467
; KNL-NEXT: retq
468
468
;
469
469
; SKX-LABEL: test_v8i32:
470
470
; SKX: # BB#0: # %entry
471
- ; SKX-NEXT: vmovdqu64 (%rdi), %ymm0
471
+ ; SKX-NEXT: vmovdqa64 (%rdi), %ymm0
472
472
; SKX-NEXT: retq
473
473
entry:
474
- %0 = load <8 x i32 >, <8 x i32 >* %V , align 16
474
+ %0 = load <8 x i32 >, <8 x i32 >* %V , align 32
475
475
ret <8 x i32 > %0
476
476
}
477
477
@@ -614,10 +614,10 @@ define <8 x float> @test_v8f32(<8 x float>* %V) {
614
614
;
615
615
; AVX-LABEL: test_v8f32:
616
616
; AVX: # BB#0: # %entry
617
- ; AVX-NEXT: vmovups (%rdi), %ymm0
617
+ ; AVX-NEXT: vmovaps (%rdi), %ymm0
618
618
; AVX-NEXT: retq
619
619
entry:
620
- %0 = load <8 x float >, <8 x float >* %V , align 16
620
+ %0 = load <8 x float >, <8 x float >* %V , align 32
621
621
ret <8 x float > %0
622
622
}
623
623
@@ -630,10 +630,10 @@ define <4 x double> @test_v4f64(<4 x double>* %V) {
630
630
;
631
631
; AVX-LABEL: test_v4f64:
632
632
; AVX: # BB#0: # %entry
633
- ; AVX-NEXT: vmovupd (%rdi), %ymm0
633
+ ; AVX-NEXT: vmovapd (%rdi), %ymm0
634
634
; AVX-NEXT: retq
635
635
entry:
636
- %0 = load <4 x double >, <4 x double >* %V , align 16
636
+ %0 = load <4 x double >, <4 x double >* %V , align 32
637
637
ret <4 x double > %0
638
638
}
639
639
@@ -692,10 +692,10 @@ define <64 x i8> @test_v64i8(<64 x i8>* %V) {
692
692
;
693
693
; SKX-LABEL: test_v64i8:
694
694
; SKX: # BB#0: # %entry
695
- ; SKX-NEXT: vmovdqu64 (%rdi), %zmm0
695
+ ; SKX-NEXT: vmovdqa64 (%rdi), %zmm0
696
696
; SKX-NEXT: retq
697
697
entry:
698
- %0 = load <64 x i8 >, <64 x i8 >* %V , align 32
698
+ %0 = load <64 x i8 >, <64 x i8 >* %V , align 64
699
699
ret <64 x i8 > %0
700
700
}
701
701
@@ -722,10 +722,10 @@ define <32 x i16> @test_v32i16(<32 x i16>* %V) {
722
722
;
723
723
; SKX-LABEL: test_v32i16:
724
724
; SKX: # BB#0: # %entry
725
- ; SKX-NEXT: vmovdqu64 (%rdi), %zmm0
725
+ ; SKX-NEXT: vmovdqa64 (%rdi), %zmm0
726
726
; SKX-NEXT: retq
727
727
entry:
728
- %0 = load <32 x i16 >, <32 x i16 >* %V , align 32
728
+ %0 = load <32 x i16 >, <32 x i16 >* %V , align 64
729
729
ret <32 x i16 > %0
730
730
}
731
731
@@ -740,16 +740,16 @@ define <16 x i32> @test_v16i32(<16 x i32>* %V) {
740
740
;
741
741
; AVXONLY-LABEL: test_v16i32:
742
742
; AVXONLY: # BB#0: # %entry
743
- ; AVXONLY-NEXT: vmovups (%rdi), %ymm0
744
- ; AVXONLY-NEXT: vmovups 32(%rdi), %ymm1
743
+ ; AVXONLY-NEXT: vmovaps (%rdi), %ymm0
744
+ ; AVXONLY-NEXT: vmovaps 32(%rdi), %ymm1
745
745
; AVXONLY-NEXT: retq
746
746
;
747
747
; AVX512-LABEL: test_v16i32:
748
748
; AVX512: # BB#0: # %entry
749
- ; AVX512-NEXT: vmovdqu64 (%rdi), %zmm0
749
+ ; AVX512-NEXT: vmovdqa64 (%rdi), %zmm0
750
750
; AVX512-NEXT: retq
751
751
entry:
752
- %0 = load <16 x i32 >, <16 x i32 >* %V , align 16
752
+ %0 = load <16 x i32 >, <16 x i32 >* %V , align 64
753
753
ret <16 x i32 > %0
754
754
}
755
755
@@ -770,10 +770,10 @@ define <8 x i64> @test_v8i64(<8 x i64>* %V) {
770
770
;
771
771
; AVX512-LABEL: test_v8i64:
772
772
; AVX512: # BB#0: # %entry
773
- ; AVX512-NEXT: vmovdqu64 (%rdi), %zmm0
773
+ ; AVX512-NEXT: vmovdqa64 (%rdi), %zmm0
774
774
; AVX512-NEXT: retq
775
775
entry:
776
- %0 = load <8 x i64 >, <8 x i64 >* %V , align 32
776
+ %0 = load <8 x i64 >, <8 x i64 >* %V , align 64
777
777
ret <8 x i64 > %0
778
778
}
779
779
@@ -894,10 +894,10 @@ define <8 x float> @test_v16f32(<8 x float>* %V) {
894
894
;
895
895
; AVX-LABEL: test_v16f32:
896
896
; AVX: # BB#0: # %entry
897
- ; AVX-NEXT: vmovups (%rdi), %ymm0
897
+ ; AVX-NEXT: vmovaps (%rdi), %ymm0
898
898
; AVX-NEXT: retq
899
899
entry:
900
- %0 = load <8 x float >, <8 x float >* %V , align 16
900
+ %0 = load <8 x float >, <8 x float >* %V , align 64
901
901
ret <8 x float > %0
902
902
}
903
903
@@ -912,16 +912,16 @@ define <8 x double> @test_v8f64(<8 x double>* %V) {
912
912
;
913
913
; AVXONLY-LABEL: test_v8f64:
914
914
; AVXONLY: # BB#0: # %entry
915
- ; AVXONLY-NEXT: vmovupd (%rdi), %ymm0
916
- ; AVXONLY-NEXT: vmovupd 32(%rdi), %ymm1
915
+ ; AVXONLY-NEXT: vmovapd (%rdi), %ymm0
916
+ ; AVXONLY-NEXT: vmovapd 32(%rdi), %ymm1
917
917
; AVXONLY-NEXT: retq
918
918
;
919
919
; AVX512-LABEL: test_v8f64:
920
920
; AVX512: # BB#0: # %entry
921
- ; AVX512-NEXT: vmovupd (%rdi), %zmm0
921
+ ; AVX512-NEXT: vmovapd (%rdi), %zmm0
922
922
; AVX512-NEXT: retq
923
923
entry:
924
- %0 = load <8 x double >, <8 x double >* %V , align 16
924
+ %0 = load <8 x double >, <8 x double >* %V , align 64
925
925
ret <8 x double > %0
926
926
}
927
927
0 commit comments