Skip to content

Releases: lucidrains/native-sparse-attention-pytorch

0.0.34

21 Feb 14:16

Choose a tag to compare

oops

0.0.33

21 Feb 14:10

Choose a tag to compare

make the fine flex block mask also aware of gqa

0.0.31

21 Feb 13:41

Choose a tag to compare

oops, think they pair up the query heads to kv heads in gqa differently

0.0.30

21 Feb 13:26

Choose a tag to compare

use enable_gqa for flex attention for the sliding windows branch

0.0.29

21 Feb 13:12

Choose a tag to compare

wire up flex fine selected attention and make sure it runs

0.0.28

21 Feb 00:34

Choose a tag to compare

Full Changelog: 0.0.27...0.0.28

0.0.27

20 Feb 18:44

Choose a tag to compare

Full Changelog: 0.0.26...0.0.27

0.0.26

20 Feb 17:16

Choose a tag to compare

Full Changelog: 0.0.24...0.0.26

0.0.25

20 Feb 17:10

Choose a tag to compare

last commit for the day, should be ready for experiments tomorrow

0.0.24

20 Feb 17:02

Choose a tag to compare

complete the fine attention masking with flex attention, not wired up