-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Bump flashinfer-python CUDA 13 #3327
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Hi @johnnynunez, thanks for contributing! We found that the change in https://github.com/apache/tvm/pull/18300/files is the only one we need for CUDA 13. Everything else is compatible with CUDA 13, including flashinfer-python. We have removed the AOT flashinfer support a while ago, so the cmake config of |
Thanks! It is because i was using thor and spark and it is compatible with cuda 13 about flashinfer with that version |
@johnnynunez Got it. We can bump it then. I am currently working on shipping our CUDA 13 python package, and will update this PR after finishing that. |
Thank you! It is the baseline for those devices |
Hi @johnnynunez, I just updated this PR and got it merged. Unfortunately FlashInfer 0.3.1 doesn't work with the latest main tvm and mlc-llm because of our recent rename of For now if you want to use FlashInfer, you will need to clone the FlashInfer GitHub repo and build it from source by following https://docs.flashinfer.ai/installation.html#python-package. |
Thor and Spark