-
Notifications
You must be signed in to change notification settings - Fork 654
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[pytorch] Allows multiple native jars to package into fat jar #1543
Conversation
Fixes deepjavalibrary#1422 Change-Id: I477d85725e84f30a7a79733b16d0224abe83f3ce
With this PR, you should be able to package multiple pytorch native library ( > 1.10.2) in a single jar |
Codecov Report
@@ Coverage Diff @@
## master #1543 +/- ##
============================================
- Coverage 72.08% 70.79% -1.30%
- Complexity 5126 5358 +232
============================================
Files 473 500 +27
Lines 21970 23375 +1405
Branches 2351 2543 +192
============================================
+ Hits 15838 16549 +711
- Misses 4925 5550 +625
- Partials 1207 1276 +69
Continue to review full report at Codecov.
|
Hi @frankfliu, Any information on when Thanks! |
It should be release together with DJL 0.17.0, we usually release new version in 45-60 days. |
@frankfliu I've got a question regarding this PR: We are using sbt to package the libraries for a fat jar, so it will include |
From design point of view, you should not bundle three of the classifier (for Mac OS, Linux, Windows) into a single jar. Each OS should have their own jars. Fat jar is used mostly on the Spark related works (mostly Linux based). DJL is using the specific OS to determine which native jar to load |
Hi @lanking520 I think the question is regarding this discussion: #1422 and more precisely this conclusion that something needs to happen to allow multiple CPU operating systems to be packaged all together: #1422 (comment) The problem with Apache Spark is that To go around this issue in Apache Spark, we package all 3 operating systems on the CPU into 1 fat jar. Obviously, the size of the package and the final Fat JAR will be larger than it should be, but it allows us to use that dependency that has a classifier on all operating systems, all machines, and all possible situations (spark-shell, spark-submit, etc.) - we do this with |
Currently we don't have a good way to avoid the Please take a look and see if that works for you. |
Fixes #1422
Change-Id: I477d85725e84f30a7a79733b16d0224abe83f3ce
Description
Brief description of what this PR is about