-
Notifications
You must be signed in to change notification settings - Fork 636
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Torch JIT breaks when memory_efficient_attention #406
Comments
Thanks for reporting :) Should be fixed in #438 |
Hello, has it been fixed right now? |
Hi, the PR was merged so it should be yes. Please let us know if you have other issues |
I'm getting this error too. |
got this error too. |
got save erros when I use torch.jit.trace ,any update? |
I think the original fix (#438) did work, but the issue was re-introduced later in #587 question to @danthe3rd , what's the purpose of two int output values |
Oh this is a regression - right. |
Does JIT support SymInt? Because the version in PT outputs |
@danthe3rd , which version of torch are you referring to? for torch 2.2.0, I see the type is
and here are when they are initialized: https://github.com/pytorch/pytorch/blob/d47f715d29d05e28b94c280f15dce097ef3dc7cb/aten/src/ATen/native/transformers/cuda/attention.cu#L978-L982 |
@danthe3rd are you referring to |
🐛 Bug
torch.jit.trace breaks with the following error:
RuntimeError: unsupported output type: int, from operator: xformers::efficient_attention_forward_generic
The output of the ops contains an int that can't be traced by JIT.
Command
To Reproduce
torch.jit.trace the module mentioned in
huggingface/diffusers#532
Expected behavior
No int output so module can be JIT traced.
The text was updated successfully, but these errors were encountered: