You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Please note that issues that do not follow the contributing guidelines are likely to be ignored.
@radna0 transformers does not support officially TPU, but I think things might work if:
you use the raw model forward API, not generate
you use the static KV cache, that works best on TPUs with Pytorch XLA.
I haven't tried this model myself, but you can try, using inspiration from this script. If it does not work, feel free to open a PR on optimum-tpu and we will try to help you.
System Info
transformers
version: 4.45.0.dev0Who can help?
No response
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
#Following this Qwen2-VL guide => https://huggingface.co/Qwen/Qwen2-VL-2B-Instruct#quickstart
Expected behavior
The model works fine when chaging
device
to"cpu"
, but stuck executing on TPUs. The model should run on TPUsThe text was updated successfully, but these errors were encountered: