Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add a flag to enable OV inference on dGPU #3503

Merged
merged 11 commits into from
May 17, 2024
Merged

Conversation

sovrasov
Copy link
Contributor

@sovrasov sovrasov commented May 15, 2024

Summary

Also, OV model outputs are now consistent with --engine.device

Performance results are contradictive, apparently we need a specially tuned inference code to get a gain from dGPU.

How to test

Checklist

  • I have added unit tests to cover my changes.​
  • I have added integration tests to cover my changes.​
  • I have ran e2e tests and there is no issues.
  • I have added the description of my changes into CHANGELOG in my target branch (e.g., CHANGELOG in develop).​
  • I have updated the documentation in my target branch accordingly (e.g., documentation in develop).
  • I have linked related issues.

License

  • I submit my code changes under the same Apache License that covers the project.
    Feel free to contact the maintainers if that's a concern.
  • I have updated the license header for each file (see an example below).
# Copyright (C) 2024 Intel Corporation
# SPDX-License-Identifier: Apache-2.0

@sovrasov sovrasov marked this pull request as draft May 15, 2024 13:25
@github-actions github-actions bot added TEST Any changes in tests DOC Improvements or additions to documentation OTX 2.0 labels May 15, 2024
Copy link

codecov bot commented May 15, 2024

Codecov Report

Attention: Patch coverage is 51.72414% with 14 lines in your changes are missing coverage. Please review.

Project coverage is 83.10%. Comparing base (4a52282) to head (1293c73).

Files Patch % Lines
src/otx/core/model/classification.py 0.00% 5 Missing ⚠️
src/otx/core/model/base.py 76.47% 4 Missing ⚠️
src/otx/core/model/instance_segmentation.py 25.00% 3 Missing ⚠️
src/otx/core/model/detection.py 33.33% 2 Missing ⚠️
Additional details and impacted files
@@             Coverage Diff             @@
##           develop    #3503      +/-   ##
===========================================
+ Coverage    83.09%   83.10%   +0.01%     
===========================================
  Files          254      254              
  Lines        25235    25236       +1     
===========================================
+ Hits         20969    20973       +4     
+ Misses        4266     4263       -3     
Flag Coverage Δ
py310 83.08% <51.72%> (+0.25%) ⬆️
py311 83.09% <51.72%> (+0.03%) ⬆️

Flags with carried forward coverage won't be shown. Click here to find out more.

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@sovrasov sovrasov changed the title Use dGPU for OV inference, if possible Add a flag to enable OV inference on dGPU May 15, 2024
@sovrasov sovrasov force-pushed the vs/ov_propagate_device branch 2 times, most recently from 9a5124a to cf52ccb Compare May 15, 2024 16:28
@sovrasov sovrasov marked this pull request as ready for review May 16, 2024 15:18
@sovrasov sovrasov merged commit 464884f into develop May 17, 2024
14 of 15 checks passed
@sovrasov sovrasov deleted the vs/ov_propagate_device branch May 17, 2024 07:47
eunwoosh pushed a commit to eunwoosh/training_extensions that referenced this pull request Jun 5, 2024
* dGPU inference for OV models

* Extract reading of hparams in OVModel

* Fix usage of get_user_config

* Fix ruff

* Add for cpu flag to ov model

* Fix missing ov core

* Fix plugin coniguration

* Add one more unit test for OVModel

* Fix imports

* Revert inf exp changes

---------

Co-authored-by: kprokofi <kirill.prokofiev@intel.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
DOC Improvements or additions to documentation TEST Any changes in tests
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants