Skip to content

[MXFP] Fix fp4 fp4 scaled_dot does not use dpas issue #4920

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

LiyangLingIntel
Copy link
Contributor

The root casue of this issue is the FP32_FP32_FP4_FP4 dot pattern is missed in DPASEngineType.
This change also adds default input precision as "tf32" when decomposing dot_scaled to dot, since dot_scaled op does not have input precision attribute.

Copy link
Contributor

@Copilot Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull Request Overview

This PR fixes an issue where FP4-FP4 scaled dot operations were not utilizing DPAS (Dot Product Accumulation Systolic) functionality. The root cause was a missing FP32_FP32_FP4_FP4 pattern in the DPAS engine type enumeration and analysis logic.

  • Added the missing FP32_FP32_FP4_FP4 DPAS engine type
  • Added detection logic for FP4-FP4 dot operations in DPAS analysis
  • Set default input precision to TF32 when decomposing scaled dot operations

Reviewed Changes

Copilot reviewed 3 out of 3 changed files in this pull request and generated no comments.

File Description
third_party/intel/include/Analysis/DPAS.h Added FP32_FP32_FP4_FP4 enum value to support FP4-FP4 DPAS operations
third_party/intel/lib/Analysis/DPAS.cpp Added detection logic for FP4-FP4 operations and removed outdated comments
third_party/intel/lib/TritonIntelGPUTransforms/DecomposeScaledBlocked.cpp Added default TF32 input precision when creating DotOp from DotScaledOp

Tip: Customize your code reviews with copilot-instructions.md. Create the file or learn how to get started.

Copy link
Contributor

@etiotto etiotto left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you add a lit test ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants