Skip to content

Commit 5c63c72

Browse files
authored
[Gluon] Implement attention kernels for d64 and d128 (#7009)
1 parent 9326a2d commit 5c63c72

File tree

2 files changed

+1177
-0
lines changed

2 files changed

+1177
-0
lines changed

Makefile

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -39,6 +39,7 @@ test-unit: all
3939
TRITON_DISABLE_LINE_INFO=0 $(PYTEST) -s python/test/unit/language/test_line_info.py
4040
# Run attention separately to avoid out of gpu memory
4141
$(PYTEST) -vs python/tutorials/06-fused-attention.py
42+
$(PYTEST) -vs python/tutorials/gluon/01-attention-forward.py
4243
TRITON_ALWAYS_COMPILE=1 TRITON_DISABLE_LINE_INFO=0 LLVM_PASS_PLUGIN_PATH=python/triton/instrumentation/libGPUInstrumentationTestLib.so \
4344
$(PYTEST) --capture=tee-sys -rfs -vvv python/test/unit/instrumentation/test_gpuhello.py
4445
$(PYTEST) -s -n $(NUM_PROCS) python/test/gluon

0 commit comments

Comments
 (0)