Skip to content

Conversation

@rampitec
Copy link
Collaborator

@rampitec rampitec commented May 6, 2025

No description provided.

Copy link
Collaborator Author

rampitec commented May 6, 2025

This stack of pull requests is managed by Graphite. Learn more about stacking.

@rampitec rampitec requested a review from arsenm May 6, 2025 20:34
@rampitec rampitec marked this pull request as ready for review May 6, 2025 20:34
@llvmbot
Copy link
Member

llvmbot commented May 6, 2025

@llvm/pr-subscribers-backend-amdgpu

Author: Stanislav Mekhanoshin (rampitec)

Changes

Full diff: https://github.com/llvm/llvm-project/pull/138756.diff

1 Files Affected:

  • (modified) llvm/lib/Target/AMDGPU/AMDGPUSubtarget.h (+6-2)
diff --git a/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.h b/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.h
index 91fe2a69bc0b7..12fd68dbea295 100644
--- a/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.h
+++ b/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.h
@@ -208,9 +208,13 @@ class AMDGPUSubtarget {
 
   bool hasFP4ConversionScaleInsts() const { return HasFP4ConversionScaleInsts; }
 
-  bool hasFP6BF6ConversionScaleInsts() const { return HasFP6BF6ConversionScaleInsts; }
+  bool hasFP6BF6ConversionScaleInsts() const {
+    return HasFP6BF6ConversionScaleInsts;
+  }
 
-  bool hasF16BF16ToFP6BF6ConversionScaleInsts() const { return HasF16BF16ToFP6BF6ConversionScaleInsts; }
+  bool hasF16BF16ToFP6BF6ConversionScaleInsts() const {
+    return HasF16BF16ToFP6BF6ConversionScaleInsts;
+  }
 
   bool hasCvtPkF16F32Inst() const { return HasCvtPkF16F32Inst; }
 

@rampitec rampitec merged commit 9ea92f0 into main May 7, 2025
15 checks passed
@rampitec rampitec deleted the users/rampitec/05-06-_amdgpu_clang-format_part_of_amdgpusubtarget.h._nfc branch May 7, 2025 02:15
GeorgeARM pushed a commit to GeorgeARM/llvm-project that referenced this pull request May 7, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants