-
Notifications
You must be signed in to change notification settings - Fork 636
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
xFormers 0.0.16
#595
Comments
can you add pytorch 2 wheels for this release? |
PyTorch 2 is not yet released, this means there is a new version every day, and we would need to release new wheels every day (because binaries are linked to a specific pt version). |
@danthe3rd would it be possible to make pip wheels available built with cuda 11.8 please? I've been trying out xformers Found that I get a decent performance boost if I manually copy over and replace the .dll files |
+1, I had to build from source for Arch Linux's CUDA 11.8 base. Is it too early for a cuda 12 wheel? |
Oh that's interesting - which GPU are you using and what did you get speedup on? For CUDA 11.8 It means you had to build torchvision and PyTorch from source? I believe the binary wheels are only for CUDA 11.6/11.7 For CUDA 12, I don't think it's supported by PyTorch. |
Hello, I am using a 4090 FE, got a ~50% speedup generating images (txt2img stable diffusion). I did not build anything from source, just a regular Then I manually replaced the installed .dll files in torch\lib with the ones from the 11.8 bin. |
Nah no torch building, I am on Arch Linux (CachyOS specifically), which builds pytorch to target the rest of their packages (Python 3.10/CUDA 11.8 at the moment): https://archlinux.org/packages/community/x86_64/python-pytorch-opt-cuda/ Right now I am just testing on a laptop RTX 2060, and the speedup is measurable but single digit %. But I am considering setting stuff up for an Ampere or Ada instance. |
this use case is actually surprisingly common (and no need to build torch from source in that case), nvidia ships cuda 11.8 with their reference docker image (I know this firsthand :)) |
We build at the moment for cuda 11.7. I believe this is compatible with cuda 11.8. |
We plan to release it as soon as #641 is fixed |
We have released |
We just released |
Todo list for
0.0.16
release:Support varying sequence lengthNot included in
0.0.16
:The text was updated successfully, but these errors were encountered: