Use the same model for GPU and CPU #1292
Unanswered
thewh1teagle
asked this question in
Q&A
Replies: 1 comment
-
@thewh1teagle Ultimately, an ONNX model will run on the ONNX Runtime. So to answer your question... you can optimize a model for CPU and then inference it on both a GPU and CPU machine. However, Olive optimizes the model for specific devices (NPU/GPU/CPU) from different hardware vendors (Nvidia, Qualcomm, etc). Therefore, you will likely get better performance by optimizing for the specific device you want to run on. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I would like to use olive to create simple single whisper medium int8 model for the app Vibe.
However it seems that olive need separate models per GPU / CPU.
I want to simplify that to single model. Is that possible?
Thanks
Beta Was this translation helpful? Give feedback.
All reactions