Enabling Blackwell support#1254
Enabling Blackwell support#1254danthe3rd merged 1 commit intofacebookresearch:mainfrom loscrossos:blackwell_support
Conversation
Signed-off-by: LosCrossos <165311345+loscrossos@users.noreply.github.com>
|
Any chance this could be prioritised? Having the same issue this PR solves and would appreciate this fix to be merged. |
|
i have a compiled library for linux python3.12 if you want to try out: https://github.com/loscrossos/xformers/releases |
|
thank you so much, I actually compilied it too in the morning (crossed my mind right after putting a comment in here). Now fine-tuning stable diffusion sdxl (dreambooth + lora). To be honest, it's not running much faster than yesterday without xformers - so I am wondering there is still some gap somewhere :D |
|
are on windows or linux? for the project i tried i also did not notice much improvement on windows. on Linux there was a performance increase. maybe there is more things to change down the line.. nevertheless this is a first step and it enables some projects to run at all. |
|
it's Win 11 by I am running WSL2 Ubuntu, the training is also running inside container. |
|
Hi, I have a qq. It looks like this PR addresses some potential incompatibility on the Blackwell GPU arch. Did you find any severe issues except for RTX5090? Our repo has some dependencies in this repository. i'm planning to use B200 machines. I'm wondering if any issues will happen, esp when using |
|
the issues are linked in the first comment. so blackwell does not run at all for those |
thx for your whl Now I'm no longer getting the error about the missing sm_120, but I'm being prompted that this version of xformers requires torch 2.7. However, I'm currently using the nightly 2.8.0.dev version. Should I ignore this prompt or downgrade torch to 2.7.0? WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for: |
|
you have CUDA 12.8 and i compiled with cuda 12.9. that is not a problem. The pytorch version is: you will have to downgrade to 2.7.0 to use my libraries. |
ok i get I'm really looking forward to your guide |
interesting I'm trying out the DeepCompressor project. Your whl file works with torch 2.8, but it just keeps giving me warnings. However, with torch 2.7, it simply won't run at all. It says it's missing GLIBCXX_3.4.31, but my libstdc++.so.6 is actually correct. |
|
you can test for it with: |
run deepcomprosser anything about cuda |
|
+1 Awaiting on this |
|
Merging this PR, although please be aware that we no longer build Flash-Attention 2 on Windows, we use Flash-Attention 3 now. |
|
@danthe3rd sorry to hijack here, but is there a way to use FA2 again on Windows? I have a multiGPU setup with Ampere, Ada and Blackwell 2.0 (8.9, 8.9 and 12.0) but out of the box it seems it doesn't build FA2 now as before. When trying to force it I get other issues, on torch nightly. With I get I have attached the file with the error |
Nvidia Blackwell cards have been out for a while bu xformers do not formally support them. This PR adds support for them by checking for CUDA12.8 and enabling capability 120. Even though higher capabilities are available in https://docs.nvidia.com/cuda/cuda-compiler-driver-nvcc/index.html#gpu-feature-list, using 120 is a conservative approach that enables all 50 series cards. This PR definitely solves #1251 and possibly #1228