Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WebNN] Better int64 integration #23831

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

[WebNN] Better int64 integration #23831

wants to merge 1 commit into from

Conversation

Honry
Copy link
Contributor

@Honry Honry commented Feb 27, 2025

This PR adds some workarounds to enable int64 support for some WebNN backends which don't support int64 data type.

  • Do not fallback ops that are specifically due to the int64 limitation.
  • Convert all int64 initializer and input values to int32 and handle potential overflow errors.
  • Register all int64 model inputs and outputs as int32 ml-tensor.
  • Handle ONNX ops that need inputs or outputs conversion between int64 and int32. e.g. ArgMax, ArgMin, Cast, etc.
  • Convert int64 output data back to int32.
  • Disallow int64 outputs as 'ml-tensor' preferredOutputLocation.

Fixed #21401

This PR adds some workarounds to enable int64 support for some WebNN
backends which don't support int64 data type.

- Do not fallback ops that are specifically due to the int64 limitation.
- Convert all int64 initializer and input values to int32 and handle
 potential overflow errors.
- Register all int64 model intputs and outputs as int32 ml-tensor.
- Handle ONNX ops that need intputs or outputs conversion between int64
and int32. e.g. ArgMax, ArgMin, Cast, etc.
- Convert int64 output data back to int32.
- Disallow int64 outputs as 'ml-tensor' preferredOutputLocation.

Fixed microsoft#21401
@Honry
Copy link
Contributor Author

Honry commented Feb 27, 2025

@fdwr, @guschmue, @egalli, PTAL, thanks!

@guschmue guschmue added the ep:WebNN WebNN execution provider label Feb 27, 2025
@guschmue
Copy link
Contributor

/azp run ONNX Runtime Web CI Pipeline,Windows GPU CI Pipeline,Linux Android Emulator QNN CI Pipeline

@guschmue
Copy link
Contributor

/azp run Linux CPU CI Pipeline,Linux CPU Minimal Build E2E CI Pipeline,Linux GPU CI Pipeline,Linux GPU TensorRT CI Pipeline, Linux OpenVINO CI Pipeline,Linux QNN CI Pipeline,MacOS CI Pipeline,Windows ARM64 QNN CI Pipeline,Windows CPU CI Pipeline

@guschmue
Copy link
Contributor

/azp run Windows GPU TensorRT CI Pipeline,onnxruntime-binary-size-checks-ci-pipeline,orttraining-linux-ci-pipeline,orttraining-linux-gpu-ci-pipeline,orttraining-ortmodule-distributed,Windows x64 QNN CI Pipeline,Big Models

Copy link

Azure Pipelines successfully started running 2 pipeline(s).

@guschmue
Copy link
Contributor

/azp run Windows GPU CUDA CI Pipeline,Windows GPU DML CI Pipeline,Windows GPU Doc Gen CI Pipeline, Win_TRT_Minimal_CUDA_Test_CI

Copy link

Azure Pipelines successfully started running 4 pipeline(s).

1 similar comment
Copy link

Azure Pipelines successfully started running 4 pipeline(s).

Copy link

Azure Pipelines successfully started running 9 pipeline(s).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ep:WebNN WebNN execution provider
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[WebNN EP] Support int64 output data type for CoreML backend
2 participants