@@ -783,7 +783,7 @@ class VPseudoUSLoadNoMask<VReg RetClass,
783
783
int EEW,
784
784
DAGOperand sewop = sew> :
785
785
Pseudo<(outs RetClass:$rd),
786
- (ins RetClass:$dest, GPRMem :$rs1, AVL:$vl, sewop:$sew,
786
+ (ins RetClass:$dest, GPRMemZeroOffset :$rs1, AVL:$vl, sewop:$sew,
787
787
vec_policy:$policy), []>,
788
788
RISCVVPseudo,
789
789
RISCVVLE</*Masked*/0, /*Strided*/0, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -800,7 +800,7 @@ class VPseudoUSLoadMask<VReg RetClass,
800
800
int EEW> :
801
801
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd),
802
802
(ins GetVRegNoV0<RetClass>.R:$passthru,
803
- GPRMem :$rs1,
803
+ GPRMemZeroOffset :$rs1,
804
804
VMaskOp:$vm, AVL:$vl, sew:$sew, vec_policy:$policy), []>,
805
805
RISCVVPseudo,
806
806
RISCVVLE</*Masked*/1, /*Strided*/0, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -818,7 +818,7 @@ class VPseudoUSLoadMask<VReg RetClass,
818
818
class VPseudoUSLoadFFNoMask<VReg RetClass,
819
819
int EEW> :
820
820
Pseudo<(outs RetClass:$rd, GPR:$vl),
821
- (ins RetClass:$dest, GPRMem :$rs1, AVL:$avl,
821
+ (ins RetClass:$dest, GPRMemZeroOffset :$rs1, AVL:$avl,
822
822
sew:$sew, vec_policy:$policy), []>,
823
823
RISCVVPseudo,
824
824
RISCVVLE</*Masked*/0, /*Strided*/0, /*FF*/1, !logtwo(EEW), VLMul> {
@@ -835,7 +835,7 @@ class VPseudoUSLoadFFMask<VReg RetClass,
835
835
int EEW> :
836
836
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd, GPR:$vl),
837
837
(ins GetVRegNoV0<RetClass>.R:$passthru,
838
- GPRMem :$rs1,
838
+ GPRMemZeroOffset :$rs1,
839
839
VMaskOp:$vm, AVL:$avl, sew:$sew, vec_policy:$policy), []>,
840
840
RISCVVPseudo,
841
841
RISCVVLE</*Masked*/1, /*Strided*/0, /*FF*/1, !logtwo(EEW), VLMul> {
@@ -853,7 +853,7 @@ class VPseudoUSLoadFFMask<VReg RetClass,
853
853
class VPseudoSLoadNoMask<VReg RetClass,
854
854
int EEW> :
855
855
Pseudo<(outs RetClass:$rd),
856
- (ins RetClass:$dest, GPRMem :$rs1, GPR:$rs2, AVL:$vl,
856
+ (ins RetClass:$dest, GPRMemZeroOffset :$rs1, GPR:$rs2, AVL:$vl,
857
857
sew:$sew, vec_policy:$policy), []>,
858
858
RISCVVPseudo,
859
859
RISCVVLE</*Masked*/0, /*Strided*/1, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -870,7 +870,7 @@ class VPseudoSLoadMask<VReg RetClass,
870
870
int EEW> :
871
871
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd),
872
872
(ins GetVRegNoV0<RetClass>.R:$passthru,
873
- GPRMem :$rs1, GPR:$rs2,
873
+ GPRMemZeroOffset :$rs1, GPR:$rs2,
874
874
VMaskOp:$vm, AVL:$vl, sew:$sew, vec_policy:$policy), []>,
875
875
RISCVVPseudo,
876
876
RISCVVLE</*Masked*/1, /*Strided*/1, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -893,7 +893,7 @@ class VPseudoILoadNoMask<VReg RetClass,
893
893
bit EarlyClobber,
894
894
bits<2> TargetConstraintType = 1> :
895
895
Pseudo<(outs RetClass:$rd),
896
- (ins RetClass:$dest, GPRMem :$rs1, IdxClass:$rs2, AVL:$vl,
896
+ (ins RetClass:$dest, GPRMemZeroOffset :$rs1, IdxClass:$rs2, AVL:$vl,
897
897
sew:$sew, vec_policy:$policy), []>,
898
898
RISCVVPseudo,
899
899
RISCVVLX</*Masked*/0, Ordered, !logtwo(EEW), VLMul, LMUL> {
@@ -916,7 +916,7 @@ class VPseudoILoadMask<VReg RetClass,
916
916
bits<2> TargetConstraintType = 1> :
917
917
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd),
918
918
(ins GetVRegNoV0<RetClass>.R:$passthru,
919
- GPRMem :$rs1, IdxClass:$rs2,
919
+ GPRMemZeroOffset :$rs1, IdxClass:$rs2,
920
920
VMaskOp:$vm, AVL:$vl, sew:$sew, vec_policy:$policy), []>,
921
921
RISCVVPseudo,
922
922
RISCVVLX</*Masked*/1, Ordered, !logtwo(EEW), VLMul, LMUL> {
@@ -936,7 +936,7 @@ class VPseudoUSStoreNoMask<VReg StClass,
936
936
int EEW,
937
937
DAGOperand sewop = sew> :
938
938
Pseudo<(outs),
939
- (ins StClass:$rd, GPRMem :$rs1, AVL:$vl, sewop:$sew), []>,
939
+ (ins StClass:$rd, GPRMemZeroOffset :$rs1, AVL:$vl, sewop:$sew), []>,
940
940
RISCVVPseudo,
941
941
RISCVVSE</*Masked*/0, /*Strided*/0, !logtwo(EEW), VLMul> {
942
942
let mayLoad = 0;
@@ -949,7 +949,7 @@ class VPseudoUSStoreNoMask<VReg StClass,
949
949
class VPseudoUSStoreMask<VReg StClass,
950
950
int EEW> :
951
951
Pseudo<(outs),
952
- (ins StClass:$rd, GPRMem :$rs1,
952
+ (ins StClass:$rd, GPRMemZeroOffset :$rs1,
953
953
VMaskOp:$vm, AVL:$vl, sew:$sew), []>,
954
954
RISCVVPseudo,
955
955
RISCVVSE</*Masked*/1, /*Strided*/0, !logtwo(EEW), VLMul> {
@@ -964,7 +964,7 @@ class VPseudoUSStoreMask<VReg StClass,
964
964
class VPseudoSStoreNoMask<VReg StClass,
965
965
int EEW> :
966
966
Pseudo<(outs),
967
- (ins StClass:$rd, GPRMem :$rs1, GPR:$rs2,
967
+ (ins StClass:$rd, GPRMemZeroOffset :$rs1, GPR:$rs2,
968
968
AVL:$vl, sew:$sew), []>,
969
969
RISCVVPseudo,
970
970
RISCVVSE</*Masked*/0, /*Strided*/1, !logtwo(EEW), VLMul> {
@@ -978,7 +978,7 @@ class VPseudoSStoreNoMask<VReg StClass,
978
978
class VPseudoSStoreMask<VReg StClass,
979
979
int EEW> :
980
980
Pseudo<(outs),
981
- (ins StClass:$rd, GPRMem :$rs1, GPR:$rs2,
981
+ (ins StClass:$rd, GPRMemZeroOffset :$rs1, GPR:$rs2,
982
982
VMaskOp:$vm, AVL:$vl, sew:$sew), []>,
983
983
RISCVVPseudo,
984
984
RISCVVSE</*Masked*/1, /*Strided*/1, !logtwo(EEW), VLMul> {
@@ -1329,7 +1329,7 @@ class VPseudoTiedBinaryNoMaskRoundingMode<VReg RetClass,
1329
1329
class VPseudoIStoreNoMask<VReg StClass, VReg IdxClass, int EEW, bits<3> LMUL,
1330
1330
bit Ordered>:
1331
1331
Pseudo<(outs),
1332
- (ins StClass:$rd, GPRMem :$rs1, IdxClass:$rs2, AVL:$vl,
1332
+ (ins StClass:$rd, GPRMemZeroOffset :$rs1, IdxClass:$rs2, AVL:$vl,
1333
1333
sew:$sew),[]>,
1334
1334
RISCVVPseudo,
1335
1335
RISCVVSX</*Masked*/0, Ordered, !logtwo(EEW), VLMul, LMUL> {
@@ -1343,7 +1343,7 @@ class VPseudoIStoreNoMask<VReg StClass, VReg IdxClass, int EEW, bits<3> LMUL,
1343
1343
class VPseudoIStoreMask<VReg StClass, VReg IdxClass, int EEW, bits<3> LMUL,
1344
1344
bit Ordered>:
1345
1345
Pseudo<(outs),
1346
- (ins StClass:$rd, GPRMem :$rs1, IdxClass:$rs2,
1346
+ (ins StClass:$rd, GPRMemZeroOffset :$rs1, IdxClass:$rs2,
1347
1347
VMaskOp:$vm, AVL:$vl, sew:$sew),[]>,
1348
1348
RISCVVPseudo,
1349
1349
RISCVVSX</*Masked*/1, Ordered, !logtwo(EEW), VLMul, LMUL> {
@@ -1598,7 +1598,7 @@ class VPseudoUSSegLoadNoMask<VReg RetClass,
1598
1598
int EEW,
1599
1599
bits<4> NF> :
1600
1600
Pseudo<(outs RetClass:$rd),
1601
- (ins RetClass:$dest, GPRMem :$rs1, AVL:$vl,
1601
+ (ins RetClass:$dest, GPRMemZeroOffset :$rs1, AVL:$vl,
1602
1602
sew:$sew, vec_policy:$policy), []>,
1603
1603
RISCVVPseudo,
1604
1604
RISCVVLSEG<NF, /*Masked*/0, /*Strided*/0, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -1615,7 +1615,7 @@ class VPseudoUSSegLoadMask<VReg RetClass,
1615
1615
int EEW,
1616
1616
bits<4> NF> :
1617
1617
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd),
1618
- (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMem :$rs1,
1618
+ (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMemZeroOffset :$rs1,
1619
1619
VMaskOp:$vm, AVL:$vl, sew:$sew, vec_policy:$policy), []>,
1620
1620
RISCVVPseudo,
1621
1621
RISCVVLSEG<NF, /*Masked*/1, /*Strided*/0, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -1634,7 +1634,7 @@ class VPseudoUSSegLoadFFNoMask<VReg RetClass,
1634
1634
int EEW,
1635
1635
bits<4> NF> :
1636
1636
Pseudo<(outs RetClass:$rd, GPR:$vl),
1637
- (ins RetClass:$dest, GPRMem :$rs1, AVL:$avl,
1637
+ (ins RetClass:$dest, GPRMemZeroOffset :$rs1, AVL:$avl,
1638
1638
sew:$sew, vec_policy:$policy), []>,
1639
1639
RISCVVPseudo,
1640
1640
RISCVVLSEG<NF, /*Masked*/0, /*Strided*/0, /*FF*/1, !logtwo(EEW), VLMul> {
@@ -1651,7 +1651,7 @@ class VPseudoUSSegLoadFFMask<VReg RetClass,
1651
1651
int EEW,
1652
1652
bits<4> NF> :
1653
1653
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd, GPR:$vl),
1654
- (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMem :$rs1,
1654
+ (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMemZeroOffset :$rs1,
1655
1655
VMaskOp:$vm, AVL:$avl, sew:$sew, vec_policy:$policy), []>,
1656
1656
RISCVVPseudo,
1657
1657
RISCVVLSEG<NF, /*Masked*/1, /*Strided*/0, /*FF*/1, !logtwo(EEW), VLMul> {
@@ -1670,7 +1670,7 @@ class VPseudoSSegLoadNoMask<VReg RetClass,
1670
1670
int EEW,
1671
1671
bits<4> NF> :
1672
1672
Pseudo<(outs RetClass:$rd),
1673
- (ins RetClass:$passthru, GPRMem :$rs1, GPR:$offset, AVL:$vl,
1673
+ (ins RetClass:$passthru, GPRMemZeroOffset :$rs1, GPR:$offset, AVL:$vl,
1674
1674
sew:$sew, vec_policy:$policy), []>,
1675
1675
RISCVVPseudo,
1676
1676
RISCVVLSEG<NF, /*Masked*/0, /*Strided*/1, /*FF*/0, !logtwo(EEW), VLMul> {
@@ -1687,7 +1687,7 @@ class VPseudoSSegLoadMask<VReg RetClass,
1687
1687
int EEW,
1688
1688
bits<4> NF> :
1689
1689
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd),
1690
- (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMem :$rs1,
1690
+ (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMemZeroOffset :$rs1,
1691
1691
GPR:$offset, VMaskOp:$vm, AVL:$vl, sew:$sew,
1692
1692
vec_policy:$policy), []>,
1693
1693
RISCVVPseudo,
@@ -1710,7 +1710,7 @@ class VPseudoISegLoadNoMask<VReg RetClass,
1710
1710
bits<4> NF,
1711
1711
bit Ordered> :
1712
1712
Pseudo<(outs RetClass:$rd),
1713
- (ins RetClass:$passthru, GPRMem :$rs1, IdxClass:$offset, AVL:$vl,
1713
+ (ins RetClass:$passthru, GPRMemZeroOffset :$rs1, IdxClass:$offset, AVL:$vl,
1714
1714
sew:$sew, vec_policy:$policy), []>,
1715
1715
RISCVVPseudo,
1716
1716
RISCVVLXSEG<NF, /*Masked*/0, Ordered, !logtwo(EEW), VLMul, LMUL> {
@@ -1732,7 +1732,7 @@ class VPseudoISegLoadMask<VReg RetClass,
1732
1732
bits<4> NF,
1733
1733
bit Ordered> :
1734
1734
Pseudo<(outs GetVRegNoV0<RetClass>.R:$rd),
1735
- (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMem :$rs1,
1735
+ (ins GetVRegNoV0<RetClass>.R:$passthru, GPRMemZeroOffset :$rs1,
1736
1736
IdxClass:$offset, VMaskOp:$vm, AVL:$vl, sew:$sew,
1737
1737
vec_policy:$policy), []>,
1738
1738
RISCVVPseudo,
@@ -1754,7 +1754,7 @@ class VPseudoUSSegStoreNoMask<VReg ValClass,
1754
1754
int EEW,
1755
1755
bits<4> NF> :
1756
1756
Pseudo<(outs),
1757
- (ins ValClass:$rd, GPRMem :$rs1, AVL:$vl, sew:$sew), []>,
1757
+ (ins ValClass:$rd, GPRMemZeroOffset :$rs1, AVL:$vl, sew:$sew), []>,
1758
1758
RISCVVPseudo,
1759
1759
RISCVVSSEG<NF, /*Masked*/0, /*Strided*/0, !logtwo(EEW), VLMul> {
1760
1760
let mayLoad = 0;
@@ -1768,7 +1768,7 @@ class VPseudoUSSegStoreMask<VReg ValClass,
1768
1768
int EEW,
1769
1769
bits<4> NF> :
1770
1770
Pseudo<(outs),
1771
- (ins ValClass:$rd, GPRMem :$rs1,
1771
+ (ins ValClass:$rd, GPRMemZeroOffset :$rs1,
1772
1772
VMaskOp:$vm, AVL:$vl, sew:$sew), []>,
1773
1773
RISCVVPseudo,
1774
1774
RISCVVSSEG<NF, /*Masked*/1, /*Strided*/0, !logtwo(EEW), VLMul> {
@@ -1784,7 +1784,7 @@ class VPseudoSSegStoreNoMask<VReg ValClass,
1784
1784
int EEW,
1785
1785
bits<4> NF> :
1786
1786
Pseudo<(outs),
1787
- (ins ValClass:$rd, GPRMem :$rs1, GPR:$offset,
1787
+ (ins ValClass:$rd, GPRMemZeroOffset :$rs1, GPR:$offset,
1788
1788
AVL:$vl, sew:$sew), []>,
1789
1789
RISCVVPseudo,
1790
1790
RISCVVSSEG<NF, /*Masked*/0, /*Strided*/1, !logtwo(EEW), VLMul> {
@@ -1799,7 +1799,7 @@ class VPseudoSSegStoreMask<VReg ValClass,
1799
1799
int EEW,
1800
1800
bits<4> NF> :
1801
1801
Pseudo<(outs),
1802
- (ins ValClass:$rd, GPRMem :$rs1, GPR: $offset,
1802
+ (ins ValClass:$rd, GPRMemZeroOffset :$rs1, GPR: $offset,
1803
1803
VMaskOp:$vm, AVL:$vl, sew:$sew), []>,
1804
1804
RISCVVPseudo,
1805
1805
RISCVVSSEG<NF, /*Masked*/1, /*Strided*/1, !logtwo(EEW), VLMul> {
@@ -1818,7 +1818,7 @@ class VPseudoISegStoreNoMask<VReg ValClass,
1818
1818
bits<4> NF,
1819
1819
bit Ordered> :
1820
1820
Pseudo<(outs),
1821
- (ins ValClass:$rd, GPRMem :$rs1, IdxClass: $index,
1821
+ (ins ValClass:$rd, GPRMemZeroOffset :$rs1, IdxClass: $index,
1822
1822
AVL:$vl, sew:$sew), []>,
1823
1823
RISCVVPseudo,
1824
1824
RISCVVSXSEG<NF, /*Masked*/0, Ordered, !logtwo(EEW), VLMul, LMUL> {
@@ -1836,7 +1836,7 @@ class VPseudoISegStoreMask<VReg ValClass,
1836
1836
bits<4> NF,
1837
1837
bit Ordered> :
1838
1838
Pseudo<(outs),
1839
- (ins ValClass:$rd, GPRMem :$rs1, IdxClass: $index,
1839
+ (ins ValClass:$rd, GPRMemZeroOffset :$rs1, IdxClass: $index,
1840
1840
VMaskOp:$vm, AVL:$vl, sew:$sew), []>,
1841
1841
RISCVVPseudo,
1842
1842
RISCVVSXSEG<NF, /*Masked*/1, Ordered, !logtwo(EEW), VLMul, LMUL> {
0 commit comments