We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 2a8f6ad commit d94f1eeCopy full SHA for d94f1ee
.github/workflows/test.yml
@@ -29,6 +29,12 @@ jobs:
29
with:
30
enable-cache: true
31
32
+ - name: Cache flash-attn build
33
+ uses: actions/cache@v4
34
+ with:
35
+ path: ~/.cache/pip
36
+ key: flash-attn-${{ runner.os }}-cuda12.5-py3.10-${{ hashFiles('uv.lock') }}
37
+
38
- name: Install dependencies
39
run: uv sync --all-groups
40
Makefile
@@ -42,7 +42,7 @@ reports:
42
.PHONY: test
43
test: reports
44
$(UV) pip install optimum-quanto
45
- $(UV) pip install flash-attn --find-links https://github.com/Dao-AILab/flash-attention/releases/expanded_assets/v2.8.3
+ $(UV) pip install flash-attn --no-build-isolation
46
PYTHONPATH=. \
47
$(UV) run pytest \
48
--cov-report xml:reports/coverage.xml \
0 commit comments