Skip to content

Flpha0830/tiny-flash-attention

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Tiny Flash Attention

A minimal Flash Attention implementation in Triton — includes both forward and backward kernels. This is just a small demo project — not intended for real-world use.

Usage

# Install dependencies using uv
uv sync

# Run all tests
uv run pytest

Requirements

  • Python ≥ 3.10
  • PyTorch ≥ 2.7.1
  • Triton ≥ 3.3.1
  • CUDA-compatible GPU

About

A minimal Flash Attention implementation in Triton

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages