@@ -690,7 +690,7 @@ static StringRef getGPUOrDefault(const Triple &TT, StringRef GPU) {
690690 return GPU;
691691
692692 // Need to default to a target with flat support for HSA.
693- if (TT.getArch () == Triple::amdgcn )
693+ if (TT.isAMDGCN () )
694694 return TT.getOS () == Triple::AMDHSA ? " generic-hsa" : " generic" ;
695695
696696 return " r600" ;
@@ -714,7 +714,7 @@ AMDGPUTargetMachine::AMDGPUTargetMachine(const Target &T, const Triple &TT,
714714 getEffectiveCodeModel(CM, CodeModel::Small), OptLevel),
715715 TLOF(createTLOF(getTargetTriple())) {
716716 initAsmInfo ();
717- if (TT.getArch () == Triple::amdgcn ) {
717+ if (TT.isAMDGCN () ) {
718718 if (getMCSubtargetInfo ()->checkFeatures (" +wavefrontsize64" ))
719719 MRI.reset (llvm::createGCNMCRegisterInfo (AMDGPUDwarfFlavour::Wave64));
720720 else if (getMCSubtargetInfo ()->checkFeatures (" +wavefrontsize32" ))
@@ -1198,8 +1198,7 @@ void AMDGPUPassConfig::addStraightLineScalarOptimizationPasses() {
11981198void AMDGPUPassConfig::addIRPasses () {
11991199 const AMDGPUTargetMachine &TM = getAMDGPUTargetMachine ();
12001200
1201- Triple::ArchType Arch = TM.getTargetTriple ().getArch ();
1202- if (RemoveIncompatibleFunctions && Arch == Triple::amdgcn)
1201+ if (RemoveIncompatibleFunctions && TM.getTargetTriple ().isAMDGCN ())
12031202 addPass (createAMDGPURemoveIncompatibleFunctionsPass (&TM));
12041203
12051204 // There is no reason to run these.
@@ -1223,7 +1222,7 @@ void AMDGPUPassConfig::addIRPasses() {
12231222 addPass (createAlwaysInlinerLegacyPass ());
12241223
12251224 // Handle uses of OpenCL image2d_t, image3d_t and sampler_t arguments.
1226- if (Arch == Triple::r600)
1225+ if (TM. getTargetTriple (). getArch () == Triple::r600)
12271226 addPass (createR600OpenCLImageTypeLoweringPass ());
12281227
12291228 // Make enqueued block runtime handles externally visible.
@@ -1242,7 +1241,7 @@ void AMDGPUPassConfig::addIRPasses() {
12421241 addPass (createInferAddressSpacesPass ());
12431242
12441243 // Run atomic optimizer before Atomic Expand
1245- if ((TM.getTargetTriple ().getArch () == Triple::amdgcn ) &&
1244+ if ((TM.getTargetTriple ().isAMDGCN () ) &&
12461245 (TM.getOptLevel () >= CodeGenOptLevel::Less) &&
12471246 (AMDGPUAtomicOptimizerStrategy != ScanOptions::None)) {
12481247 addPass (createAMDGPUAtomicOptimizerPass (AMDGPUAtomicOptimizerStrategy));
@@ -1265,7 +1264,7 @@ void AMDGPUPassConfig::addIRPasses() {
12651264 }));
12661265 }
12671266
1268- if (TM.getTargetTriple ().getArch () == Triple::amdgcn ) {
1267+ if (TM.getTargetTriple ().isAMDGCN () ) {
12691268 // TODO: May want to move later or split into an early and late one.
12701269 addPass (createAMDGPUCodeGenPreparePass ());
12711270 }
@@ -1295,17 +1294,16 @@ void AMDGPUPassConfig::addIRPasses() {
12951294}
12961295
12971296void AMDGPUPassConfig::addCodeGenPrepare () {
1298- if (TM->getTargetTriple ().getArch () == Triple::amdgcn ) {
1297+ if (TM->getTargetTriple ().isAMDGCN () ) {
12991298 // FIXME: This pass adds 2 hacky attributes that can be replaced with an
13001299 // analysis, and should be removed.
13011300 addPass (createAMDGPUAnnotateKernelFeaturesPass ());
13021301 }
13031302
1304- if (TM->getTargetTriple ().getArch () == Triple::amdgcn &&
1305- EnableLowerKernelArguments)
1303+ if (TM->getTargetTriple ().isAMDGCN () && EnableLowerKernelArguments)
13061304 addPass (createAMDGPULowerKernelArgumentsPass ());
13071305
1308- if (TM->getTargetTriple ().getArch () == Triple::amdgcn ) {
1306+ if (TM->getTargetTriple ().isAMDGCN () ) {
13091307 // This lowering has been placed after codegenprepare to take advantage of
13101308 // address mode matching (which is why it isn't put with the LDS lowerings).
13111309 // It could be placed anywhere before uniformity annotations (an analysis
0 commit comments