Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Default layout inference pass #704

Open
spacycoder opened this issue Sep 26, 2024 · 5 comments
Open

Default layout inference pass #704

spacycoder opened this issue Sep 26, 2024 · 5 comments

Comments

@spacycoder
Copy link

spacycoder commented Sep 26, 2024

Hi, I get these warnings when running an int8 quantized model on I.MX 8M Plus using the vx delegate:

W [HandleLayoutInfer:332]Op 162: default layout inference pass.
W [HandleLayoutInfer:332]Op 56: default layout inference pass.

What does these warnings mean?

The model performs worse on NPU than CPU which I'm guessing might be related to these warning

@sergiomsilva
Copy link

Hi @spacycoder , have you figured out? I'm having the same warnings and performance issues. It seems related to matmul operations inside a custom layer I have.

@spacycoder
Copy link
Author

No I haven't, but I also suspect it has something to do with matmul and possibly reshape ops

@sunshinemyson
Copy link
Contributor

@spacycoder @sergiomsilva ,

"default layout inference pass" means that operation is not supported by layout inference framework - additional transpose operation may insert before/after your operation.

It could happen if the operation is customized op or reshape operation.

Thanks

@BralSLA
Copy link

BralSLA commented Dec 5, 2024

@spacycoder @sergiomsilva ,

"default layout inference pass" means that operation is not supported by layout inference framework - additional transpose operation may insert before/after your operation.

It could happen if the operation is customized op or reshape operation.

Thanks

@sunshinemyson Currently facing the same issue. Is this something that can be solved using a different version of the vx_delegate? Would like to not rely on changing the model architecture if possible. Thanks.

@hgaiser
Copy link

hgaiser commented Dec 16, 2024

I got the exact same problem, a bunch of these warnings and the performance is worse compared to just CPU. Has anyone figured out what causes this?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants