Skip to content
Discussion options

You must be logged in to vote

The JIT package should be available to use for any cuda + torch versions:
You can try:

pip install flashinfer-python

We plan to forsake the AOT wheels, and only keeps the JIT package, because in the long term binary size would exceed limits of pypi wheel size even if we apply for growth.

Currently user can try:

python -m flashinfer.aot

to compile all kernels ahead-of-time.

We are building the artifactory so that user can try (it's not ready yet but should be available soon @yongwww ):

python -m flashinfer --download all

to download all pre-built binaries.

Replies: 2 comments

Comment options

You must be logged in to vote
0 replies
Answer selected by cdunn95
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants