-
Notifications
You must be signed in to change notification settings - Fork 85
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Default layout inference pass #704
Comments
Hi @spacycoder , have you figured out? I'm having the same warnings and performance issues. It seems related to matmul operations inside a custom layer I have. |
No I haven't, but I also suspect it has something to do with matmul and possibly reshape ops |
"default layout inference pass" means that operation is not supported by layout inference framework - additional transpose operation may insert before/after your operation. It could happen if the operation is customized op or reshape operation. Thanks |
@sunshinemyson Currently facing the same issue. Is this something that can be solved using a different version of the vx_delegate? Would like to not rely on changing the model architecture if possible. Thanks. |
I got the exact same problem, a bunch of these warnings and the performance is worse compared to just CPU. Has anyone figured out what causes this? |
Hi, I get these warnings when running an int8 quantized model on I.MX 8M Plus using the vx delegate:
W [HandleLayoutInfer:332]Op 162: default layout inference pass.
W [HandleLayoutInfer:332]Op 56: default layout inference pass.
What does these warnings mean?
The model performs worse on NPU than CPU which I'm guessing might be related to these warning
The text was updated successfully, but these errors were encountered: