|
25 | 25 | "--iree-preprocessing-pass-pipeline=builtin.module(iree-preprocessing-transpose-convolution-pipeline, iree-global-opt-raise-special-ops, util.func(iree-preprocessing-pad-to-intrinsics, iree-linalg-ext-pad-attention{pad-to-multiple-of=0,64,0,32,0}))", |
26 | 26 | ], |
27 | 27 | "punet": [ |
28 | | - "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-raise-special-ops, iree-flow-canonicalize), iree-preprocessing-transpose-convolution-pipeline, util.func(iree-preprocessing-pad-to-intrinsics, iree-preprocessing-generalize-linalg-matmul-experimental))" |
| 28 | + "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-raise-special-ops, iree-dispatch-creation-canonicalize), iree-preprocessing-transpose-convolution-pipeline, iree-preprocessing-pad-to-intrinsics, util.func(iree-preprocessing-generalize-linalg-matmul-experimental))" |
29 | 29 | ], |
30 | 30 | "vae_preprocess": [ |
31 | | - "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-raise-special-ops, iree-flow-canonicalize), iree-preprocessing-transpose-convolution-pipeline, util.func(iree-preprocessing-pad-to-intrinsics, iree-preprocessing-generalize-linalg-matmul-experimental))" |
| 31 | + "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-raise-special-ops, iree-dispatch-creation-canonicalize), iree-preprocessing-transpose-convolution-pipeline, iree-preprocessing-pad-to-intrinsics, util.func(iree-preprocessing-generalize-linalg-matmul-experimental))" |
32 | 32 | ], |
33 | 33 | "preprocess_default": [ |
34 | | - "--iree-preprocessing-pass-pipeline=builtin.module(iree-preprocessing-transpose-convolution-pipeline, util.func(iree-preprocessing-pad-to-intrinsics))", |
| 34 | + "--iree-preprocessing-pass-pipeline=builtin.module(iree-preprocessing-transpose-convolution-pipeline, iree-preprocessing-pad-to-intrinsics)", |
35 | 35 | ], |
36 | 36 | "unet": [ |
37 | | - "--iree-flow-enable-aggressive-fusion", |
| 37 | + "--iree-dispatch-creation-enable-aggressive-fusion", |
38 | 38 | "--iree-opt-aggressively-propagate-transposes=true", |
39 | 39 | "--iree-codegen-llvmgpu-use-vector-distribution=true", |
40 | 40 | "--iree-opt-outer-dim-concat=true", |
|
43 | 43 | "--iree-vm-target-truncate-unsupported-floats", |
44 | 44 | ], |
45 | 45 | "clip": [ |
46 | | - "--iree-flow-enable-aggressive-fusion", |
47 | | - "--iree-flow-enable-fuse-horizontal-contractions=true", |
| 46 | + "--iree-dispatch-creation-enable-aggressive-fusion", |
| 47 | + "--iree-dispatch-creation-enable-fuse-horizontal-contractions=true", |
48 | 48 | "--iree-opt-aggressively-propagate-transposes=true", |
49 | 49 | "--iree-opt-outer-dim-concat=true", |
50 | 50 | "--iree-rocm-waves-per-eu=2", |
51 | 51 | "--iree-codegen-llvmgpu-use-vector-distribution=true", |
52 | 52 | ], |
53 | 53 | "vae": [ |
54 | | - "--iree-flow-enable-aggressive-fusion", |
55 | | - "--iree-flow-enable-fuse-horizontal-contractions", |
| 54 | + "--iree-dispatch-creation-enable-aggressive-fusion", |
| 55 | + "--iree-dispatch-creation-enable-fuse-horizontal-contractions", |
56 | 56 | "--iree-opt-aggressively-propagate-transposes=true", |
57 | 57 | "--iree-codegen-llvmgpu-use-vector-distribution=true", |
58 | 58 | "--iree-opt-data-tiling=false", |
|
70 | 70 | "--iree-opt-data-tiling=false", |
71 | 71 | "--iree-opt-const-eval=false", |
72 | 72 | "--iree-opt-aggressively-propagate-transposes=true", |
73 | | - "--iree-flow-enable-aggressive-fusion", |
| 73 | + "--iree-dispatch-creation-enable-aggressive-fusion", |
74 | 74 | "--iree-codegen-gpu-native-math-precision=true", |
75 | 75 | "--iree-codegen-llvmgpu-use-vector-distribution=true", |
76 | 76 | ], |
|
80 | 80 | "--iree-codegen-llvmgpu-enable-transform-dialect-jit=false", |
81 | 81 | ], |
82 | 82 | "punet": [ |
83 | | - "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-raise-special-ops, iree-flow-canonicalize), iree-preprocessing-transpose-convolution-pipeline, iree-preprocessing-pad-to-intrinsics, util.func(iree-preprocessing-generalize-linalg-matmul-experimental))" |
84 | | - "--iree-flow-enable-fuse-horizontal-contractions=true", |
| 83 | + "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-raise-special-ops, iree-dispatch-creation-canonicalize), iree-preprocessing-transpose-convolution-pipeline, iree-preprocessing-pad-to-intrinsics, util.func(iree-preprocessing-generalize-linalg-matmul-experimental))" |
| 84 | + "--iree-dispatch-creation-enable-fuse-horizontal-contractions=true", |
85 | 85 | "--iree-codegen-llvmgpu-enable-transform-dialect-jit=false", |
86 | 86 | ], |
87 | 87 | "preprocess_default": [ |
88 | | - "--iree-preprocessing-pass-pipeline=builtin.module(iree-preprocessing-transpose-convolution-pipeline, iree-global-opt-raise-special-ops, util.func(iree-preprocessing-pad-to-intrinsics))", |
| 88 | + "--iree-preprocessing-pass-pipeline=builtin.module(iree-preprocessing-transpose-convolution-pipeline, iree-global-opt-raise-special-ops, iree-preprocessing-pad-to-intrinsics)", |
89 | 89 | "--iree-codegen-llvmgpu-enable-transform-dialect-jit=false", |
90 | 90 | ], |
91 | 91 | "unet": [""], |
|
98 | 98 | "--iree-llvmcpu-target-cpu=znver4", |
99 | 99 | "--iree-opt-const-eval=false", |
100 | 100 | "--iree-llvmcpu-enable-ukernels=mmt4d,pack,unpack", |
101 | | - "--iree-flow-collapse-reduction-dims", |
| 101 | + "--iree-dispatch-creation-collapse-reduction-dims", |
102 | 102 | "--iree-opt-const-expr-max-size-increase-threshold=1000000000000000", |
103 | | - "--iree-flow-enable-fuse-padding-into-linalg-consumer-ops", |
| 103 | + "--iree-dispatch-creation-enable-fuse-padding-into-linalg-consumer-ops", |
104 | 104 | ], |
105 | 105 | "bf16": [ |
106 | 106 | "--iree-preprocessing-pass-pipeline=builtin.module(util.func(iree-global-opt-demote-contraction-inputs-to-bf16))", |
@@ -288,15 +288,15 @@ def compile_to_vmfb( |
288 | 288 | "--iree-stream-resource-max-allocation-size=" + max_alloc, |
289 | 289 | "--iree-stream-resource-index-bits=64", |
290 | 290 | "--iree-vm-target-index-bits=64", |
291 | | - "--iree-flow-inline-constants-max-byte-length=1", |
| 291 | + "--iree-dispatch-creation-inline-constants-max-byte-length=1", |
292 | 292 | ] |
293 | 293 | ) |
294 | 294 | device = "vulkan-spirv" |
295 | 295 | elif device in ["rocm", "hip"]: |
296 | 296 | flags.extend( |
297 | 297 | [ |
298 | 298 | "--iree-hal-target-backends=rocm", |
299 | | - "--iree-rocm-target-chip=" + target_triple, |
| 299 | + "--iree-hip-target=" + target_triple, |
300 | 300 | "--iree-vm-bytecode-module-output-format=flatbuffer-binary", |
301 | 301 | ] |
302 | 302 | ) |
|
0 commit comments