-
Notifications
You must be signed in to change notification settings - Fork 15.4k
Reland "[flang] Inline hlfir.dot_product. (#123143)" #123385
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This reverts commit afc43a7. +Fixed declaration of hlfir::genExtentsVector().
|
@llvm/pr-subscribers-flang-fir-hlfir Author: Slava Zakharin (vzakhari) ChangesThis reverts commit afc43a7. Patch is 30.92 KiB, truncated to 20.00 KiB below, full version: https://github.com/llvm/llvm-project/pull/123385.diff 4 Files Affected:
diff --git a/flang/include/flang/Optimizer/Builder/HLFIRTools.h b/flang/include/flang/Optimizer/Builder/HLFIRTools.h
index 6e85b8f4ddf86e..0684ad0f926ec9 100644
--- a/flang/include/flang/Optimizer/Builder/HLFIRTools.h
+++ b/flang/include/flang/Optimizer/Builder/HLFIRTools.h
@@ -513,6 +513,12 @@ genTypeAndKindConvert(mlir::Location loc, fir::FirOpBuilder &builder,
Entity loadElementAt(mlir::Location loc, fir::FirOpBuilder &builder,
Entity entity, mlir::ValueRange oneBasedIndices);
+/// Return a vector of extents for the given entity.
+/// The function creates new operations, but tries to clean-up
+/// after itself.
+llvm::SmallVector<mlir::Value, Fortran::common::maxRank>
+genExtentsVector(mlir::Location loc, fir::FirOpBuilder &builder, Entity entity);
+
} // namespace hlfir
#endif // FORTRAN_OPTIMIZER_BUILDER_HLFIRTOOLS_H
diff --git a/flang/lib/Optimizer/Builder/HLFIRTools.cpp b/flang/lib/Optimizer/Builder/HLFIRTools.cpp
index 5e5d0bbd681326..f71adf123511d5 100644
--- a/flang/lib/Optimizer/Builder/HLFIRTools.cpp
+++ b/flang/lib/Optimizer/Builder/HLFIRTools.cpp
@@ -1421,3 +1421,15 @@ hlfir::Entity hlfir::loadElementAt(mlir::Location loc,
return loadTrivialScalar(loc, builder,
getElementAt(loc, builder, entity, oneBasedIndices));
}
+
+llvm::SmallVector<mlir::Value, Fortran::common::maxRank>
+hlfir::genExtentsVector(mlir::Location loc, fir::FirOpBuilder &builder,
+ hlfir::Entity entity) {
+ entity = hlfir::derefPointersAndAllocatables(loc, builder, entity);
+ mlir::Value shape = hlfir::genShape(loc, builder, entity);
+ llvm::SmallVector<mlir::Value, Fortran::common::maxRank> extents =
+ hlfir::getExplicitExtentsFromShape(shape, builder);
+ if (shape.getUses().empty())
+ shape.getDefiningOp()->erase();
+ return extents;
+}
diff --git a/flang/lib/Optimizer/HLFIR/Transforms/SimplifyHLFIRIntrinsics.cpp b/flang/lib/Optimizer/HLFIR/Transforms/SimplifyHLFIRIntrinsics.cpp
index 0fe3620b7f1ae3..fe7ae0eeed3cc3 100644
--- a/flang/lib/Optimizer/HLFIR/Transforms/SimplifyHLFIRIntrinsics.cpp
+++ b/flang/lib/Optimizer/HLFIR/Transforms/SimplifyHLFIRIntrinsics.cpp
@@ -37,6 +37,79 @@ static llvm::cl::opt<bool> forceMatmulAsElemental(
namespace {
+// Helper class to generate operations related to computing
+// product of values.
+class ProductFactory {
+public:
+ ProductFactory(mlir::Location loc, fir::FirOpBuilder &builder)
+ : loc(loc), builder(builder) {}
+
+ // Generate an update of the inner product value:
+ // acc += v1 * v2, OR
+ // acc += CONJ(v1) * v2, OR
+ // acc ||= v1 && v2
+ //
+ // CONJ parameter specifies whether the first complex product argument
+ // needs to be conjugated.
+ template <bool CONJ = false>
+ mlir::Value genAccumulateProduct(mlir::Value acc, mlir::Value v1,
+ mlir::Value v2) {
+ mlir::Type resultType = acc.getType();
+ acc = castToProductType(acc, resultType);
+ v1 = castToProductType(v1, resultType);
+ v2 = castToProductType(v2, resultType);
+ mlir::Value result;
+ if (mlir::isa<mlir::FloatType>(resultType)) {
+ result = builder.create<mlir::arith::AddFOp>(
+ loc, acc, builder.create<mlir::arith::MulFOp>(loc, v1, v2));
+ } else if (mlir::isa<mlir::ComplexType>(resultType)) {
+ if constexpr (CONJ)
+ result = fir::IntrinsicLibrary{builder, loc}.genConjg(resultType, v1);
+ else
+ result = v1;
+
+ result = builder.create<fir::AddcOp>(
+ loc, acc, builder.create<fir::MulcOp>(loc, result, v2));
+ } else if (mlir::isa<mlir::IntegerType>(resultType)) {
+ result = builder.create<mlir::arith::AddIOp>(
+ loc, acc, builder.create<mlir::arith::MulIOp>(loc, v1, v2));
+ } else if (mlir::isa<fir::LogicalType>(resultType)) {
+ result = builder.create<mlir::arith::OrIOp>(
+ loc, acc, builder.create<mlir::arith::AndIOp>(loc, v1, v2));
+ } else {
+ llvm_unreachable("unsupported type");
+ }
+
+ return builder.createConvert(loc, resultType, result);
+ }
+
+private:
+ mlir::Location loc;
+ fir::FirOpBuilder &builder;
+
+ mlir::Value castToProductType(mlir::Value value, mlir::Type type) {
+ if (mlir::isa<fir::LogicalType>(type))
+ return builder.createConvert(loc, builder.getIntegerType(1), value);
+
+ // TODO: the multiplications/additions by/of zero resulting from
+ // complex * real are optimized by LLVM under -fno-signed-zeros
+ // -fno-honor-nans.
+ // We can make them disappear by default if we:
+ // * either expand the complex multiplication into real
+ // operations, OR
+ // * set nnan nsz fast-math flags to the complex operations.
+ if (fir::isa_complex(type) && !fir::isa_complex(value.getType())) {
+ mlir::Value zeroCmplx = fir::factory::createZeroValue(builder, loc, type);
+ fir::factory::Complex helper(builder, loc);
+ mlir::Type partType = helper.getComplexPartType(type);
+ return helper.insertComplexPart(zeroCmplx,
+ castToProductType(value, partType),
+ /*isImagPart=*/false);
+ }
+ return builder.createConvert(loc, type, value);
+ }
+};
+
class TransposeAsElementalConversion
: public mlir::OpRewritePattern<hlfir::TransposeOp> {
public:
@@ -90,11 +163,8 @@ class TransposeAsElementalConversion
static mlir::Value genResultShape(mlir::Location loc,
fir::FirOpBuilder &builder,
hlfir::Entity array) {
- mlir::Value inShape = hlfir::genShape(loc, builder, array);
- llvm::SmallVector<mlir::Value> inExtents =
- hlfir::getExplicitExtentsFromShape(inShape, builder);
- if (inShape.getUses().empty())
- inShape.getDefiningOp()->erase();
+ llvm::SmallVector<mlir::Value, 2> inExtents =
+ hlfir::genExtentsVector(loc, builder, array);
// transpose indices
assert(inExtents.size() == 2 && "checked in TransposeOp::validate");
@@ -137,7 +207,7 @@ class SumAsElementalConversion : public mlir::OpRewritePattern<hlfir::SumOp> {
mlir::Value resultShape, dimExtent;
llvm::SmallVector<mlir::Value> arrayExtents;
if (isTotalReduction)
- arrayExtents = genArrayExtents(loc, builder, array);
+ arrayExtents = hlfir::genExtentsVector(loc, builder, array);
else
std::tie(resultShape, dimExtent) =
genResultShapeForPartialReduction(loc, builder, array, dimVal);
@@ -163,7 +233,8 @@ class SumAsElementalConversion : public mlir::OpRewritePattern<hlfir::SumOp> {
// If DIM is not present, do total reduction.
// Initial value for the reduction.
- mlir::Value reductionInitValue = genInitValue(loc, builder, elementType);
+ mlir::Value reductionInitValue =
+ fir::factory::createZeroValue(builder, loc, elementType);
// The reduction loop may be unordered if FastMathFlags::reassoc
// transformations are allowed. The integer reduction is always
@@ -264,17 +335,6 @@ class SumAsElementalConversion : public mlir::OpRewritePattern<hlfir::SumOp> {
}
private:
- static llvm::SmallVector<mlir::Value>
- genArrayExtents(mlir::Location loc, fir::FirOpBuilder &builder,
- hlfir::Entity array) {
- mlir::Value inShape = hlfir::genShape(loc, builder, array);
- llvm::SmallVector<mlir::Value> inExtents =
- hlfir::getExplicitExtentsFromShape(inShape, builder);
- if (inShape.getUses().empty())
- inShape.getDefiningOp()->erase();
- return inExtents;
- }
-
// Return fir.shape specifying the shape of the result
// of a SUM reduction with DIM=dimVal. The second return value
// is the extent of the DIM dimension.
@@ -283,7 +343,7 @@ class SumAsElementalConversion : public mlir::OpRewritePattern<hlfir::SumOp> {
fir::FirOpBuilder &builder,
hlfir::Entity array, int64_t dimVal) {
llvm::SmallVector<mlir::Value> inExtents =
- genArrayExtents(loc, builder, array);
+ hlfir::genExtentsVector(loc, builder, array);
assert(dimVal > 0 && dimVal <= static_cast<int64_t>(inExtents.size()) &&
"DIM must be present and a positive constant not exceeding "
"the array's rank");
@@ -293,26 +353,6 @@ class SumAsElementalConversion : public mlir::OpRewritePattern<hlfir::SumOp> {
return {builder.create<fir::ShapeOp>(loc, inExtents), dimExtent};
}
- // Generate the initial value for a SUM reduction with the given
- // data type.
- static mlir::Value genInitValue(mlir::Location loc,
- fir::FirOpBuilder &builder,
- mlir::Type elementType) {
- if (auto ty = mlir::dyn_cast<mlir::FloatType>(elementType)) {
- const llvm::fltSemantics &sem = ty.getFloatSemantics();
- return builder.createRealConstant(loc, elementType,
- llvm::APFloat::getZero(sem));
- } else if (auto ty = mlir::dyn_cast<mlir::ComplexType>(elementType)) {
- mlir::Value initValue = genInitValue(loc, builder, ty.getElementType());
- return fir::factory::Complex{builder, loc}.createComplex(ty, initValue,
- initValue);
- } else if (mlir::isa<mlir::IntegerType>(elementType)) {
- return builder.createIntegerConstant(loc, elementType, 0);
- }
-
- llvm_unreachable("unsupported SUM reduction type");
- }
-
// Generate scalar addition of the two values (of the same data type).
static mlir::Value genScalarAdd(mlir::Location loc,
fir::FirOpBuilder &builder,
@@ -570,16 +610,10 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
static std::tuple<mlir::Value, mlir::Value>
genResultShape(mlir::Location loc, fir::FirOpBuilder &builder,
hlfir::Entity input1, hlfir::Entity input2) {
- mlir::Value input1Shape = hlfir::genShape(loc, builder, input1);
- llvm::SmallVector<mlir::Value> input1Extents =
- hlfir::getExplicitExtentsFromShape(input1Shape, builder);
- if (input1Shape.getUses().empty())
- input1Shape.getDefiningOp()->erase();
- mlir::Value input2Shape = hlfir::genShape(loc, builder, input2);
- llvm::SmallVector<mlir::Value> input2Extents =
- hlfir::getExplicitExtentsFromShape(input2Shape, builder);
- if (input2Shape.getUses().empty())
- input2Shape.getDefiningOp()->erase();
+ llvm::SmallVector<mlir::Value, 2> input1Extents =
+ hlfir::genExtentsVector(loc, builder, input1);
+ llvm::SmallVector<mlir::Value, 2> input2Extents =
+ hlfir::genExtentsVector(loc, builder, input2);
llvm::SmallVector<mlir::Value, 2> newExtents;
mlir::Value innerProduct1Extent, innerProduct2Extent;
@@ -627,60 +661,6 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
innerProductExtent[0]};
}
- static mlir::Value castToProductType(mlir::Location loc,
- fir::FirOpBuilder &builder,
- mlir::Value value, mlir::Type type) {
- if (mlir::isa<fir::LogicalType>(type))
- return builder.createConvert(loc, builder.getIntegerType(1), value);
-
- // TODO: the multiplications/additions by/of zero resulting from
- // complex * real are optimized by LLVM under -fno-signed-zeros
- // -fno-honor-nans.
- // We can make them disappear by default if we:
- // * either expand the complex multiplication into real
- // operations, OR
- // * set nnan nsz fast-math flags to the complex operations.
- if (fir::isa_complex(type) && !fir::isa_complex(value.getType())) {
- mlir::Value zeroCmplx = fir::factory::createZeroValue(builder, loc, type);
- fir::factory::Complex helper(builder, loc);
- mlir::Type partType = helper.getComplexPartType(type);
- return helper.insertComplexPart(
- zeroCmplx, castToProductType(loc, builder, value, partType),
- /*isImagPart=*/false);
- }
- return builder.createConvert(loc, type, value);
- }
-
- // Generate an update of the inner product value:
- // acc += v1 * v2, OR
- // acc ||= v1 && v2
- static mlir::Value genAccumulateProduct(mlir::Location loc,
- fir::FirOpBuilder &builder,
- mlir::Type resultType,
- mlir::Value acc, mlir::Value v1,
- mlir::Value v2) {
- acc = castToProductType(loc, builder, acc, resultType);
- v1 = castToProductType(loc, builder, v1, resultType);
- v2 = castToProductType(loc, builder, v2, resultType);
- mlir::Value result;
- if (mlir::isa<mlir::FloatType>(resultType))
- result = builder.create<mlir::arith::AddFOp>(
- loc, acc, builder.create<mlir::arith::MulFOp>(loc, v1, v2));
- else if (mlir::isa<mlir::ComplexType>(resultType))
- result = builder.create<fir::AddcOp>(
- loc, acc, builder.create<fir::MulcOp>(loc, v1, v2));
- else if (mlir::isa<mlir::IntegerType>(resultType))
- result = builder.create<mlir::arith::AddIOp>(
- loc, acc, builder.create<mlir::arith::MulIOp>(loc, v1, v2));
- else if (mlir::isa<fir::LogicalType>(resultType))
- result = builder.create<mlir::arith::OrIOp>(
- loc, acc, builder.create<mlir::arith::AndIOp>(loc, v1, v2));
- else
- llvm_unreachable("unsupported type");
-
- return builder.createConvert(loc, resultType, result);
- }
-
static mlir::LogicalResult
genContiguousMatmul(mlir::Location loc, fir::FirOpBuilder &builder,
hlfir::Entity result, mlir::Value resultShape,
@@ -748,9 +728,9 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
hlfir::loadElementAt(loc, builder, lhs, {I, K});
hlfir::Entity rhsElementValue =
hlfir::loadElementAt(loc, builder, rhs, {K, J});
- mlir::Value productValue = genAccumulateProduct(
- loc, builder, resultElementType, resultElementValue,
- lhsElementValue, rhsElementValue);
+ mlir::Value productValue =
+ ProductFactory{loc, builder}.genAccumulateProduct(
+ resultElementValue, lhsElementValue, rhsElementValue);
builder.create<hlfir::AssignOp>(loc, productValue, resultElement);
return {};
};
@@ -785,9 +765,9 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
hlfir::loadElementAt(loc, builder, lhs, {J, K});
hlfir::Entity rhsElementValue =
hlfir::loadElementAt(loc, builder, rhs, {K});
- mlir::Value productValue = genAccumulateProduct(
- loc, builder, resultElementType, resultElementValue,
- lhsElementValue, rhsElementValue);
+ mlir::Value productValue =
+ ProductFactory{loc, builder}.genAccumulateProduct(
+ resultElementValue, lhsElementValue, rhsElementValue);
builder.create<hlfir::AssignOp>(loc, productValue, resultElement);
return {};
};
@@ -817,9 +797,9 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
hlfir::loadElementAt(loc, builder, lhs, {K});
hlfir::Entity rhsElementValue =
hlfir::loadElementAt(loc, builder, rhs, {K, J});
- mlir::Value productValue = genAccumulateProduct(
- loc, builder, resultElementType, resultElementValue,
- lhsElementValue, rhsElementValue);
+ mlir::Value productValue =
+ ProductFactory{loc, builder}.genAccumulateProduct(
+ resultElementValue, lhsElementValue, rhsElementValue);
builder.create<hlfir::AssignOp>(loc, productValue, resultElement);
return {};
};
@@ -885,9 +865,9 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
hlfir::loadElementAt(loc, builder, lhs, lhsIndices);
hlfir::Entity rhsElementValue =
hlfir::loadElementAt(loc, builder, rhs, rhsIndices);
- mlir::Value productValue = genAccumulateProduct(
- loc, builder, resultElementType, reductionArgs[0], lhsElementValue,
- rhsElementValue);
+ mlir::Value productValue =
+ ProductFactory{loc, builder}.genAccumulateProduct(
+ reductionArgs[0], lhsElementValue, rhsElementValue);
return {productValue};
};
llvm::SmallVector<mlir::Value, 1> innerProductValue =
@@ -904,6 +884,73 @@ class MatmulConversion : public mlir::OpRewritePattern<Op> {
}
};
+class DotProductConversion
+ : public mlir::OpRewritePattern<hlfir::DotProductOp> {
+public:
+ using mlir::OpRewritePattern<hlfir::DotProductOp>::OpRewritePattern;
+
+ llvm::LogicalResult
+ matchAndRewrite(hlfir::DotProductOp product,
+ mlir::PatternRewriter &rewriter) const override {
+ hlfir::Entity op = hlfir::Entity{product};
+ if (!op.isScalar())
+ return rewriter.notifyMatchFailure(product, "produces non-scalar result");
+
+ mlir::Location loc = product.getLoc();
+ fir::FirOpBuilder builder{rewriter, product.getOperation()};
+ hlfir::Entity lhs = hlfir::Entity{product.getLhs()};
+ hlfir::Entity rhs = hlfir::Entity{product.getRhs()};
+ mlir::Type resultElementType = product.getType();
+ bool isUnordered = mlir::isa<mlir::IntegerType>(resultElementType) ||
+ mlir::isa<fir::LogicalType>(resultElementType) ||
+ static_cast<bool>(builder.getFastMathFlags() &
+ mlir::arith::FastMathFlags::reassoc);
+
+ mlir::Value extent = genProductExtent(loc, builder, lhs, rhs);
+
+ auto genBody = [&](mlir::Location loc, fir::FirOpBuilder &builder,
+ mlir::ValueRange oneBasedIndices,
+ mlir::ValueRange reductionArgs)
+ -> llvm::SmallVector<mlir::Value, 1> {
+ hlfir::Entity lhsElementValue =
+ hlfir::loadElementAt(loc, builder, lhs, oneBasedIndices);
+ hlfir::Entity rhsElementValue =
+ hlfir::loadElementAt(loc, builder, rhs, oneBasedIndices);
+ mlir::Value productValue =
+ ProductFactory{loc, builder}.genAccumulateProduct</*CONJ=*/true>(
+ reductionArgs[0], lhsElementValue, rhsElementValue);
+ return {productValue};
+ };
+
+ mlir::Value initValue =
+ fir::factory::createZeroValue(builder, loc, resultElementType);
+
+ llvm::SmallVector<mlir::Value, 1> result = hlfir::genLoopNestWithReductions(
+ loc, builder, {extent},
+ /*reductionInits=*/{initValue}, genBody, isUnordered);
+
+ rewriter.replaceOp(product, result[0]);
+ return mlir::success();
+ }
+
+private:
+ static mlir::Value genProductExtent(mlir::Location loc,
+ fir::FirOpBuilder &builder,
+ hlfir::Entity input1,
+ hlfir::Entity input2) {
+ llvm::SmallVector<mlir::Value, 1> input1Extents =
+ hlfir::genExtentsVector(loc, builder, input1);
+ llvm::SmallVector<mlir::Value, 1> input2Extents =
+ hlfir::genExtentsVector(loc, builder, input2);
+
+ assert(input1Extents.size() == 1 && input2Extents.size() == 1 &&
+ "hlfir.dot_product arguments must be vectors");
+ llvm::SmallVector<mlir::Value, 1> extent =
+ fir::factory::deduceOptimalExtents(input1Extents, input2Extents);
+ return extent[0];
+ }
+};
+
class SimplifyHLFIRIntrinsics
: public hlfir::impl::SimplifyHLFIRIntrinsicsBase<SimplifyHLFIRIntrinsics> {
public:
@@ -939,6 +986,8 @@ class SimplifyHLFIRIntrinsics
if (forceMatmulAsElemental || this->allowNewSideEffects)
patterns.insert<MatmulConversion<hlfir::MatmulOp>>(context);
+ patterns.insert<DotProductConversion>(context);
+
if (mlir::failed(mlir::applyPatternsGreedily(
getOperation(), std::move(patterns), config))) {
mlir::emitError(getOperation()->getLoc(),
diff --git a/flang/test/HLFIR/simplify-hlfir-intrinsics-dotproduct.fir b/flang/test/HLFIR/simplify-hlfir-intrinsics-dotproduct.fir
new file mode 100644
index 00000000000000..f59b1422dbc849
--- /dev/null
+++ b/flang/test/HLFIR/simplify-hlfir-intrinsics-dotproduct.fir
@@ -0,0 +1,144 @@
+// Test hlfir.dot_product...
[truncated]
|
|
@preames thank you for reporting the issue. |
|
You don't need to re-review for a reland as long as the fix is relatively obvious and described in the commit message. You should make sure to repeat the original commit message at the end of the message body - remember that this is the change folks are going to run across in e.g. blame. |
|
I wanted to run the pre-commit pipeline. Thanks for taking a look! |
This reverts commit afc43a7.
+Fixed declaration of hlfir::genExtentsVector().
Some good results for induct2, where dot_product is applied
to a vector of unknow size and a known 3-element vector:
the inlining ends up generating a 3-iteration loop, which
is then fully unrolled. With late FIR simplification
it is not happening even when the simplified intrinsics
implementation is inlined by LLVM (because the loop bounds
are not known).
This change just follows the current approach to expose
the loops for later worksharing application.