Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sync with apache/incubator-tvm 6/15/2020 #116

Merged
merged 360 commits into from
Jun 16, 2020
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
360 commits
Select commit Hold shift + click to select a range
349b1eb
[TFLITE]Select op support for tflite frontend (#5486)
siju-samuel May 7, 2020
a5cfce7
[FRONTEND][TFLite] Fully connected op conversion made in sync with T…
May 7, 2020
6b2323e
[TOPI][Winograd] Optimization of Conv2d Winograd algorithm on Tensor …
wsl-inspur May 7, 2020
25dfd69
Cache PrimExpr instead of raw pointers in bound analyzer (#5533)
May 7, 2020
cfb41e6
fix a few bugs with shape inference and types in the onnx importer (#…
May 7, 2020
ce4d49a
[Frontend][TFLite] ADD_N operator (#5474)
maheshambule May 7, 2020
5e36e5f
[WEB][RUNTIME] TVM WebAssembly JS Runtime (#5506)
tqchen May 7, 2020
132017d
[RELAY][ONNX]ReduceLogSumExp Operator support (#5453)
siju-samuel May 7, 2020
476623a
[RPC][BUGFIX] Fix remote device sync (#5538)
tqchen May 7, 2020
1c8b943
[Refactor][std::string --> String] IRModule is updated with String (#…
May 7, 2020
5040831
[RUNTIME] Store nullptr PackedFunc as nullptr for better error propag…
tqchen May 8, 2020
fca80dd
[Relay-TFLite] FP32 and Quantized Object Detection Model (#5479)
anijain2305 May 8, 2020
4988478
Changes to cpp_rpc to make it work on Android (+ Hexagon offloading) …
May 8, 2020
612b828
Add Onnx Pad v11 (#5539)
May 8, 2020
a420710
fix restructured text (#5541)
tmoreau89 May 8, 2020
9754024
[CRT]fix to reduce RAM size during loading model (#5507)
siju-samuel May 8, 2020
72ade90
Load platform specific lib for tvmdsoop instead of only so (#5542)
tobegit3hub May 8, 2020
4f8d7ef
[RPC] Improve RPCServer AsyncIO support. (#5544)
tqchen May 8, 2020
f623acd
[Rust] Add first stage of updating and rewriting Rust bindings. (#5526)
jroesch May 8, 2020
74a687d
[TE] Fix MakeLoopNest for warp memory (#5382)
roastduck May 9, 2020
7630339
[TIR][Printer] text format printer considering future parsing use (#5…
spectrometerHBH May 9, 2020
57e9178
[Optimization] Warp level reduction support for CUDA (#5498)
wpan11nv May 9, 2020
c1cb6de
A clone of test/python/unittest/test_runtime_micro.py, however (#5546)
tom-gall May 9, 2020
37b3c97
[CI] Install wasmtime for WebAssembly tests (#5494)
kazum May 9, 2020
fb7c648
Apparently, ONNX Conv with no 'pads' defaults to zero padding (#5548)
May 9, 2020
46a00ef
[WEB] WebGPU support (#5545)
tqchen May 9, 2020
8e21d89
[TOPI][RELAY][TENSORFLOW]Math ops added (#5502)
siju-samuel May 11, 2020
76a3069
[RUNTIME] Hexagon driver for offloading kernels to simulator (#5492)
May 11, 2020
7003426
[LINT] clang-format the h,cc,m files. (#5557)
tqchen May 11, 2020
b346536
[BYOC, MergeComposite] Add additional check before re-using the cache…
masahi May 11, 2020
5eeb654
[WEB] Setup lint, doc, test (#5556)
tqchen May 11, 2020
fc14b92
[CI] Update ci-cpu to bionic (#5555)
tqchen May 11, 2020
bc67bd2
[CI] Update ci-cpu to bionic (#5554)
tqchen May 11, 2020
f79e38f
[Fix] Fix conv2d alter op for arm cpu (#5532)
icemelon May 11, 2020
fbbf7e6
[FRONTEND]onnx, mxnet, pytorch mathops added (#5561)
siju-samuel May 11, 2020
6fdf2d8
Fix topi test for tensorcore (#5563)
May 11, 2020
5f1e17d
[Refactor][std::string --> String] IR is updated with String (#5547)
May 11, 2020
25150cb
[DOCKER] Fix vulkansdk in the ci-gpu (#5566)
tqchen May 11, 2020
a22d3b2
[CI] reintroduce docker stage for wasm tests (#5565)
tqchen May 11, 2020
205092b
[CI] Update ci-lint to use the latest image that contains clang-forma…
tqchen May 12, 2020
c12a177
[DOCKER] Add clang-format and nodejs to ci-lint (#5567)
tqchen May 12, 2020
0346399
[TARGET] Phase out WebGL (#5570)
tqchen May 12, 2020
12b0fee
[LINT] Enable clang-format. (#5572)
tqchen May 12, 2020
392e8e7
[CI] Update the ci-gpu to the lastest build with the new vulkansdk. (…
tqchen May 12, 2020
b98ddd6
[Relay] enable blocking format in x86 conv2d and fold scale axis (#5357)
Menooker May 12, 2020
3fb5940
[CI] Fix clang-format error (#5577)
tqchen May 12, 2020
c80c416
Allow ubuntu_install_darknet.sh to work in both 18.04 and 16.04 (#5574)
May 12, 2020
a8b2e6b
[PYTORCH]expand bug fix (#5576)
siju-samuel May 13, 2020
c011e78
[CI] Enable llvm-11 and llvm-10 in build tests, recover webdocs. (#5579)
tqchen May 13, 2020
38e21c5
[PYTORCH] Support max_pool2d_with_indices (#5549)
May 13, 2020
e0a16fd
[Relay] Fixed bug in attribute parsing for pool layers. (#5582)
jwfromm May 13, 2020
e23c04e
[RELAY][TF] Support symbolic newshape for Reshape (#5429)
lixiaoquan May 13, 2020
293d719
Add prim::device op (#5584)
Xuxue1 May 13, 2020
1bc37f3
Fix the runtime raise error (#5586)
huochaitiantang May 13, 2020
4d148f4
[RELAY][Convert Layout] Specify additional layouts in convert layout …
lhutton1 May 13, 2020
b45fa57
Add a quantized conv2 unit test for the tflite front-end (#5558)
giuseros May 13, 2020
05d5ccd
[Relay][Transform] Safe check added for Merge Composite (#5562)
May 13, 2020
d98eeaa
[MXNET]abs, round, reciprocal, sign, softsign, hard_sigmoid (#5587)
siju-samuel May 13, 2020
7802479
[Hexagon] One more fix for concurrency count (#5589)
May 14, 2020
4341abe
Fix JSON graph dumping. (#5591)
areusch May 14, 2020
ed2d2dc
[DOCS] Improve document in reflection (#5593)
liangfu May 14, 2020
3c94d51
Overestimate binary size for microTVM compiled binaries. (#5590)
areusch May 14, 2020
1c7941c
[TFLite Runtime] Fix bug and re-enable RPC execution test (#5436)
michalpiszczek May 15, 2020
7c4270b
[Relay][VM] Memory planner (part 1) (#5144)
jroesch May 15, 2020
d61732f
Add ostream formatters for TargetPtr/TargetVal. (#5592)
areusch May 15, 2020
994c3d9
Pattern Language, Matcher, Rewriter, and Function Paritioner (#5231)
May 15, 2020
b8aebb1
[Reduction] Fix cross thread redunction (#5551)
wpan11nv May 15, 2020
1e00918
Fix TVMArray layout on device (#5599)
areusch May 15, 2020
bd4cb8b
[LLVM] Represent alignment information in LLVM IR (#5598)
May 15, 2020
ba58db9
Add debug mode to tempdir() (#5581)
areusch May 15, 2020
2f63c8a
[PYTORCH]ImplicitTensorToNum support added (#5603)
siju-samuel May 15, 2020
892274a
[PYTORCH]Matmul fix for batch_matmul (#5604)
siju-samuel May 15, 2020
38d0448
fix rpc server bug on VTA (#5607)
tmoreau89 May 16, 2020
52ca6c6
[REFACTOR][IR] Streamline ir/op Registry (#5609)
tqchen May 17, 2020
1c60e71
[TFLITE]GATHER_ND (#5508)
dhruvaray May 18, 2020
a7e6a08
[CUDA] Fix codegen for warp shuffle intrinsics (#5606)
roastduck May 18, 2020
7499c91
Fix a typo. (#5611)
littlefish0123 May 18, 2020
ed63f03
fix pattern topological order (#5612)
May 18, 2020
a641641
[BYOC] Remove kCompiler attr from external functions (#5615)
lhutton1 May 18, 2020
078e3b5
[Relay]Improve Shape Func handling for Tuple inputs (#5467)
kevinthesun May 18, 2020
22b54bd
[Relay][Refactor][std::string --> String] Relay updated with String (…
May 18, 2020
6a8d14b
[KERAS]Global MaxPool3d and AvgPool3d support (#5098)
siju-samuel May 19, 2020
a0d1b76
[IOS] Fix build error of iOS RPC (#5621)
kazum May 19, 2020
67575e0
Fix three typos (#5620)
littlefish0123 May 19, 2020
1b0b71d
[Frontend][Tensorflow] Gather nd bug fix for one dim support in tenso…
May 19, 2020
c6ae0e1
[MXNET]MaxPool3d and AvgPool3d Ops support added (#5614)
siju-samuel May 20, 2020
f3f66d2
[PYTORCH]ReflectionPad2d op (#5624)
siju-samuel May 20, 2020
84b1a26
[BYOC][MergeComposite] if root->args[i] isn't a CallNode, then Donwca…
windclarion May 20, 2020
da9c8b6
[DOCS] Move the api docs to the api subfolder (#5626)
tqchen May 20, 2020
1d5f63d
[RELAY][BYOC] Fix the creation of tuple of tuples in PartitionGraph (…
manupak May 20, 2020
b34f96a
[NODE][PASS] Introduce config to PassContext. (#5631)
tqchen May 20, 2020
d757502
another cmake fix (#5630)
tmoreau89 May 20, 2020
3963a46
Fix typo in test script (#5635)
vinx13 May 20, 2020
f06a3db
Label Pattern Partitions (#5627)
May 21, 2020
1e494d2
[RELAY][PYTORCH]Resize3d, Upsample3d op support (#5633)
siju-samuel May 21, 2020
069a591
[TUTORIAL]TFLite QNN Tutorial (#5595)
siju-samuel May 21, 2020
4087560
Extend AttrPattern to support CallNode and FunctionNode attributes (#…
May 21, 2020
93446e6
[DOCS] Fix the QNN TFLite tutorial build (#5641)
tqchen May 21, 2020
ca1ade8
[RUNTIME][VULKAN] Seg fault in WorkspacePool's destructor (#5632) (#5…
samwyi May 21, 2020
33b9b63
[PYTORCH]Padding support (#5638)
siju-samuel May 21, 2020
4b5ec64
Remove unnecessary print (#5642)
comaniac May 21, 2020
caacf21
[CI] Allow CI_PYTEST_ADD_OPTIONS to be unbound. (#5644)
tqchen May 21, 2020
8c7f09a
[Runtime] Introduce runtime::Array (#5585)
junrushao May 21, 2020
ee41673
[CI] Add log check to the sphinx gallery docs (#5643)
tqchen May 22, 2020
855b02b
[RELAY][BYOC] Preserve type information in Merge Composite (#5640)
mbaret May 22, 2020
58a8047
Add a check Callback to the Pattern Paritioner (#5646)
May 22, 2020
6a174f7
[Relay, Topi][OP] Correlation (#5628)
vinx13 May 22, 2020
23f3057
HG: Commit message of changeset 6281661. (#5622)
hlu1 May 22, 2020
68f6fb1
[AutoTVM] Update XGBoost verbosity option (#5649)
tqchen May 22, 2020
2a5d7cb
[RUNTIME] Resolve constexpr issue in debug mode. (#5651)
tqchen May 22, 2020
6285923
µtvm debug improvements (#5648)
areusch May 22, 2020
e312fcb
[REFACTOR][IR] Migrate IRModule ObjectRef to not-null (#5654)
May 22, 2020
08bbbbf
Upgrade XGBoost to latest (#5658)
hcho3 May 23, 2020
6d6889a
Increase bss section size. (#5660)
areusch May 23, 2020
727f2bd
[PatternLang] Convert PatternGrouper to do pre-order, non-recursive a…
May 23, 2020
8ba1434
[Relay,Topi][OP] affine_grid and grid_sample (#5657)
vinx13 May 23, 2020
7c2f729
[TIR][BUILD] Remove buffer params from pass config. (#5652)
tqchen May 23, 2020
6e1c282
handle likely in IRMutatorWithAnalyzer (#5665)
roastduck May 25, 2020
8dd1981
[TOPI] Improve CUDA softmax scheduling (#5600)
wpan11nv May 25, 2020
2e6f759
[Relay][Op]Support symbolic TopK, Ones, Zeros and Full (#5459)
kevinthesun May 26, 2020
0645ddc
[PYTHON] Add buffer name when creating tensor bindings (#5670)
tqchen May 26, 2020
e1f8dc3
[REFACTOR][TIR][API-Change] Migrate BuildConfig to PassContext. (#5668)
tqchen May 26, 2020
0572864
[Doc] Misc doc fix (#5672)
FrozenGene May 26, 2020
5197650
[C++ RPC] Fix C++ RPC build problem on Linux (#5671)
FrozenGene May 26, 2020
67fa7f6
enable amd_apu device on vulkan target (#5659)
mei-ye May 26, 2020
88ca82e
[AutoTVM][TOPI] AutoTVM incorrect measurement (#5511)
cchung100m May 26, 2020
12725c6
[POC][PatternLang]Remove constants from partitioned functions (#5663)
May 26, 2020
e2cd522
[TF] Support TupleWrapper as direct ancestor of control flow ops (#5639)
lixiaoquan May 26, 2020
bd14bd7
add tvm.micro pydoc to sphinx (#5661)
areusch May 26, 2020
18e01de
add a check for null function attributes (#5674)
May 26, 2020
f538f3e
[BYOC] Pattern Language MergeComposite (#5656)
comaniac May 26, 2020
15651bc
add a testcase for #5674 (#5677)
May 27, 2020
6d0d74e
Call previous excepthook in tvm_excepthook. (#5675)
notoraptor May 27, 2020
115ee90
Fix the shift column for scale_shift_nchw and scale_shift_nhwc in C t…
tobegit3hub May 27, 2020
5d0f22d
[Bugfix] Fix Python debugger segfaults with TVM built with LLVM (#5685)
junrushao May 27, 2020
1eb0677
[DOC] Improve Pattern Language Docs (#5676)
comaniac May 28, 2020
a7ab91e
[TFLITE]Quantize & Dequantize op (#5394)
siju-samuel May 28, 2020
e03458b
[TIR][REFACTOR] std::string -> String Migration in TIR nodes (#5596)
cchung100m May 28, 2020
3389e00
[PatternLang] Add ConstantPattern (#5689)
comaniac May 28, 2020
e92e3ff
[PYTORCH]Minor bug fixes (#5683)
siju-samuel May 29, 2020
bbf6289
[Relay] Fix dataflow_pattern.rewrite() hang if Match in IR (#5680)
lixiaoquan May 29, 2020
8ba91d5
[RELAY] Fix segfault in pretty print when ObjectRef is null (#5681)
lhutton1 May 29, 2020
e75d9fe
[REFACTOR][RELAY] move fallback_device to config (#5690)
zhiics May 29, 2020
82903b8
@zhiics -> PPMC (#5692)
tqchen May 29, 2020
6a4bebc
[COMMUNITY] @masahi -> PPMC (#5691)
tqchen May 29, 2020
92f4fd1
Support more dtypes for TVMDSOOp (#5694)
tobegit3hub May 29, 2020
22761ab
[ONNX]LpPool Support added (#5696)
siju-samuel May 29, 2020
81990f7
In memory_plan, check if value is not None, instead of just checking …
notoraptor May 29, 2020
8151305
[PatternLang]Conditionally Embedding Constants in Partitioned Functio…
May 30, 2020
887da51
[ONNX] Skip ADD inside Gemm op when vector is zero (#5697)
cbalint13 May 30, 2020
db46cc8
[BYOC] Support Tuple Output in C/DNNL Codegen (#5701)
comaniac May 30, 2020
d06ed8e
[REFACTOR][RELAY] Replace build_config with PassContext (#5698)
zhiics May 30, 2020
6e74866
[PYTORCH]floor_divide support for squeezenet (#5702)
siju-samuel May 30, 2020
aca92e6
[AutoTVM][TOPI] Fix bifrost spatial packing conv2d auto tune (#5684)
cchung100m May 31, 2020
519c4ad
[Arith] ExtendedEuclidean merge impl to int_operator (#5625)
Jun 1, 2020
dc9cee0
fix typo: anchor windoes should be anchor windows (#5706)
randxie Jun 1, 2020
3c718f1
[REFACTOR][PY] relay.op.Op -> tvm.ir.Op (#5705)
tqchen Jun 1, 2020
aab95e8
[PatternLang] Simplify Pattern API Implementations (#5703)
comaniac Jun 2, 2020
f848dff
[PYTORCH]ReplicationPad support added (#5708)
siju-samuel Jun 2, 2020
6344d06
Remove deprecated opengl files (#5711)
tqchen Jun 2, 2020
4606693
Remove opengl runtime and cmake (#5712)
tqchen Jun 2, 2020
c8ab3f6
[BUGFIX][CRT] Fix Compilation Error in CRT (#5713)
liangfu Jun 2, 2020
c29f3b7
Rename tvm_dso_op to libtvm_dso_op (#5714)
tobegit3hub Jun 2, 2020
2b4b60e
[Object] Unify StrMapNode and MapNode (#5687)
junrushao Jun 2, 2020
a274dd4
[MXNET]Softmin, trunc op support added (#5715)
siju-samuel Jun 3, 2020
e587a13
Avoid downloading when TOPHUB_LOCATION is NONE (#5720)
lixiaoquan Jun 3, 2020
932e434
[Object][FFI] Introduce runtime::String::CanConvertFrom (#5718)
junrushao Jun 3, 2020
ccd455c
[Object] Restore the StrMap behavior in JSON/SHash/SEqual (#5719)
junrushao Jun 3, 2020
5049211
Fix generating types like float44 and float88 (#5722)
abergeron Jun 3, 2020
9b7a754
[ONNX]ReduceL1, ReduceL2, ReduceSumSquare, ReduceLogSum ops added (#5…
siju-samuel Jun 4, 2020
5aca76f
[TENSORFLOW]StatefulPartitionedCall/PartitionedCall Ops support added…
deepakbabel23 Jun 4, 2020
5733295
[AutoTVM, Relay] Clear compile engine after task extraction (#5724)
vinx13 Jun 4, 2020
e769544
Fix runtime::String backward compatibility in JSON (#5725)
junrushao Jun 4, 2020
ddad27f
codegen llvm: move nvptx-specific intrinsic handling into codegen_nvp…
t-vi Jun 4, 2020
6436080
[TOPI,RELAY][TFLITE] Sparse to dense operator (#5447)
dhruvaray Jun 4, 2020
fd4c26c
[Frontend][TFLite] Add parser support for shape and range (#5329)
dhruvaray Jun 4, 2020
bdede15
[REFACTOR] Separate ArgTypeCode from DLDataTypeCode (#5730)
tqchen Jun 4, 2020
9ecfc74
[ONNX]MaxRoiPool, Mod & Xor op support added (#5729)
siju-samuel Jun 5, 2020
ad1122a
ROCm: Add warp shuffles and enable reductions (#5727)
t-vi Jun 5, 2020
50d79ff
Change 'delete's in Relay VM Instruction dtor to 'delete[]'s (#5735)
akosik-anyvision Jun 5, 2020
6996fb6
Fix reshape usage in ARM Winograd (#5732)
comaniac Jun 5, 2020
0078773
[TEST] Fix flaky topi/tests/python/test_topi_pooling.py:test_adaptive…
tqchen Jun 5, 2020
4879749
Fix the values for test_fmod since it fails way too often otherwise (…
abergeron Jun 5, 2020
6c7a32b
fix small bug about dense_grad (#5695)
handar423 Jun 6, 2020
b6278bf
[REFACTOR][ARITH] Remove legacy compute_expr.h (#5738)
tqchen Jun 6, 2020
6790a19
Add some docs on downstream consistency (#5742)
junrushao Jun 6, 2020
2fab9c1
sequential cpp test (#5745)
zhiics Jun 7, 2020
4fefa1d
[REFACTOR][TE][TIR] Call::Halide => ProducerLoad, DSL/TIR decouple. (…
tqchen Jun 7, 2020
da3947c
Don't add cast for TF batch norm when type isn't changing (#5731)
Jun 8, 2020
b93338e
[ARITH][BACKPORT-0.6] fix a min/max simplify bug (#5749)
xqdan Jun 9, 2020
a6eba1a
[TOPI][Relay][OP] support dynamic NMS(Non Maximum Suppression), symbo…
yongwww Jun 9, 2020
e4e1914
Update dmlc_tvm_commit_id.txt
Jun 9, 2020
349819f
Update TRT Integration to reflect upstream changes
Jun 9, 2020
bdf03dd
Sync submodules
Jun 9, 2020
c20e34c
Fix jenkinsfile
Jun 9, 2020
95144ee
git-clang-format against origin/dev instead of origin/master
Jun 9, 2020
bcfa727
Fix formatting.
Jun 9, 2020
2238600
Remove is_empty in export_lib (used for old trt)
Jun 10, 2020
719c374
Disable test_forward_qnn_mobilenet_v2_net
Jun 11, 2020
8f6bf2a
Add Scatter to Topi/Relay/ONNX via hybrid script (#5619)
Jun 9, 2020
b6bbed0
[Minor][Test] Clean WASM environment before build (#5759)
junrushao Jun 10, 2020
eaaefd4
[Bugfix] Fix reshape (#5739)
comaniac Jun 10, 2020
f4e8cf1
[REFACTOR][TIR] Provide->ProducerStore, Realize->ProducerRealize. (#5…
tqchen Jun 10, 2020
f12956b
[Rust] Second stage of Rust Refactor (#5527)
jroesch Jun 10, 2020
e7a3b38
[topi] block sparse dense on cuda (#5746)
ceruleangu Jun 10, 2020
827e103
[Relay] Fix for recursive let (#5757)
icemelon Jun 11, 2020
53dcdb0
[TOPI][RELAY][PYTORCH]Conv3d_transpose op support added (#5737)
siju-samuel Jun 11, 2020
14157dd
Fix gelu in PyTorch frontend, tighten numerical checks (#5763)
t-vi Jun 11, 2020
56dfec2
Add ShapePattern and DataTypePattern (#5760)
Jun 11, 2020
629ad8d
Make batch matrix multiplication on GPU tunable (#5752)
t-vi Jun 11, 2020
f0b6eae
[TIR][REFACTOR][API-Change] Migrate the tvm/tir/expr.h to construct s…
tqchen Jun 11, 2020
1309af6
[TIR][REFACTOR][API-Change] Migrate tir/stmt.h to use constructor. (#…
tqchen Jun 11, 2020
0e0a69c
[Frontend][TensorFlow] Improve Control Flow and TensorArray (#5699)
kevinthesun Jun 12, 2020
f0dc96b
[DOC][FIX] Fix some typos in git-clang-format.sh (#5786)
zhiics Jun 12, 2020
f6d8444
fix #5686: remove a overstrict assert in MakeAllreduce (#5686) (#5785)
majiang31312 Jun 12, 2020
5c39c30
[RUNTIME] Add compile_shared option to linux compile utility fn (#5751)
wrongtest-intellif Jun 12, 2020
b2e172c
[REFACTOR][API-Change] Migrate all Object construction to constructor…
tqchen Jun 12, 2020
79d6d65
[Topi] pass-by-value -> pass-by-const-reference (#5783)
hlu1 Jun 12, 2020
9a38788
[topi][relay] Add operation gather to relay. (#5716)
notoraptor Jun 12, 2020
9895e24
[CODEGEN][CONTRIB] CoreML codegen (#5634)
kazum Jun 12, 2020
e86a7a7
fix calibration pass to support multiple functions (#5768)
seanlatias Jun 12, 2020
63c2e30
[cmake] update vulkan rules (#5777)
antinucleon Jun 12, 2020
71be16c
Add ignore storage_order attribute to onnx pooling parser. (#5781)
jwfromm Jun 12, 2020
0ec274f
[BYOC][FIX] Infer types in MergeComposite (#5766)
mbaret Jun 12, 2020
826816c
[FRONTEND]Darknet support batch size for yolo (#5688)
siju-samuel Jun 12, 2020
ca3cb75
Update dmlc_tvm_commid_id.txt
Jun 12, 2020
f627892
Skip tflite test_forward_mediapipe_hand_landmark
Jun 12, 2020
2283275
Increase stack limit for failing tflite tests. Skip TF tests which re…
Jun 15, 2020
6f63123
[PYTORCH]aten::norm support added (#5776)
siju-samuel Jun 12, 2020
79721f8
[TENSORFLOW]Conv3d Transpose OP added (#5775)
siju-samuel Jun 12, 2020
15709c2
[TF] Support symbolic inputs of Fill (#5762)
lixiaoquan Jun 12, 2020
5522ad6
[COMMUNITY] @wpan11nv -> Reviewer (#5790)
tqchen Jun 12, 2020
534eccf
Edit onnx parser to infer values in post order (#5755)
Jun 12, 2020
ae745ea
[TIR][REFACTOR] Cleanup unused classes (#5789)
tqchen Jun 13, 2020
a9aa8ac
Fix tf parser (#5794)
kevinthesun Jun 13, 2020
e21351c
support aten::type_as in the pytorch frontend (#5787)
randxie Jun 13, 2020
9eb29b6
[TIR][REFACTIR] Update TIR nodes std::string->String. (#5793)
tqchen Jun 13, 2020
ca14048
[TEST] Temporary disable fp16 type_as test for PyTorch Frontend (#5799)
tqchen Jun 13, 2020
29e2ec7
[ONNX] Skip multiply with 1.0f constant for GEMM import (#5800)
cbalint13 Jun 14, 2020
34a581f
[TIR][REFACTOR] Add tir prefix to type keys (#5802)
tqchen Jun 14, 2020
f250700
[QUANTIZE] Add config switch for nn.dense layer type. (#5801)
cbalint13 Jun 14, 2020
33fcf79
[topi] fix sparse dense schedule on cuda (#5803)
ceruleangu Jun 14, 2020
8e18755
Allow RPCWrappedFunc to rewrite runtime::String as std::string (#5796)
junrushao Jun 14, 2020
2ca5680
[topi] fix strategy for sparse dense cuda (#5782)
antinucleon Jun 15, 2020
eecc5d2
[CI] Move cpu-only frontend tests to a CPU stage (#5807)
tqchen Jun 15, 2020
89160b9
[MXNET]conv3d and conv3d_transpose addedx (#5814)
siju-samuel Jun 15, 2020
ffb4004
Pin hand landmark network to version 0.7.4. (#5813)
leandron Jun 15, 2020
bc5a78d
[CI] Limit number of threads in all jobs (#5815)
tqchen Jun 15, 2020
13290ab
Update dmlc_tvm_commit_id.txt
Jun 15, 2020
dbb760c
Disable tensorflow.test_forward_sdd because stack limit of 100mb is e…
Jun 15, 2020
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
7 changes: 4 additions & 3 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -2,10 +2,11 @@
__pycache__/
*.py[cod]
*$py.class

*.S
# C extensions
*.so

*.ll
.npm
# Distribution / packaging
.Python
env/
Expand Down Expand Up @@ -224,7 +225,7 @@ Pipfile.lock
# conda package artifacts
conda/Dockerfile.cuda*
conda/pkg

.node_repl_history
# nix files
.envrc
*.nix
Expand Down
4 changes: 2 additions & 2 deletions 3rdparty/bfloat16/bfloat16.cc
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,7 @@
==============================================================================*/

#include <tvm/runtime/c_runtime_api.h>

#include <cstddef>
#include <cstdint>

Expand Down Expand Up @@ -50,8 +51,7 @@ void BFloat16ToFloat(const uint16_t* src, float* dst, size_t size) {
#endif
}

void BFloat16Add(const uint16_t* a, const uint16_t* b, uint16_t* dst,
size_t size) {
void BFloat16Add(const uint16_t* a, const uint16_t* b, uint16_t* dst, size_t size) {
float a_f, b_f;
BFloat16ToFloat(a, &a_f, 1);
BFloat16ToFloat(b, &b_f, 1);
Expand Down
17 changes: 7 additions & 10 deletions 3rdparty/cma/cma.h
Original file line number Diff line number Diff line change
Expand Up @@ -27,20 +27,17 @@
#ifndef VTA_DE10_NANO_KERNEL_MODULE_CMA_H_
#define VTA_DE10_NANO_KERNEL_MODULE_CMA_H_


/* Should be defined in settings.mk file */
#ifndef CMA_IOCTL_MAGIC
#define CMA_IOCTL_MAGIC 0xf2
#define CMA_IOCTL_MAGIC 0xf2
#endif

#define CMA_ALLOC_CACHED _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 1, 4)
#define CMA_ALLOC_NONCACHED _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 2, 4)
#define CMA_FREE _IOC(_IOC_WRITE, CMA_IOCTL_MAGIC, 3, 4)
#define CMA_GET_PHY_ADDR _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 4, 4)
#define CMA_GET_SIZE _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 5, 4)

#define CMA_ALLOC_CACHED _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 1, 4)
#define CMA_ALLOC_NONCACHED _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 2, 4)
#define CMA_FREE _IOC(_IOC_WRITE, CMA_IOCTL_MAGIC, 3, 4)
#define CMA_GET_PHY_ADDR _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 4, 4)
#define CMA_GET_SIZE _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 5, 4)

#define CMA_IOCTL_MAXNR 5

#define CMA_IOCTL_MAXNR 5

#endif // VTA_DE10_NANO_KERNEL_MODULE_CMA_H_
62 changes: 28 additions & 34 deletions 3rdparty/cma/cma_api_impl.h
Original file line number Diff line number Diff line change
Expand Up @@ -30,48 +30,47 @@
* \brief Application layer implementation for contigous memory allocation.
*/

#include <errno.h>
#include <fcntl.h>
#include <stdio.h>
#include <stdlib.h>
#include <fcntl.h>
#include <unistd.h>
#include <errno.h>
#include <string.h>
#include <sys/types.h>
#include <sys/ioctl.h>
#include <sys/mman.h>
#include <sys/types.h>
#include <unistd.h>

#include "cma_api.h"

#ifndef CMA_IOCTL_MAGIC
#define CMA_IOCTL_MAGIC 0xf2
#define CMA_IOCTL_MAGIC 0xf2
#endif

#define CMA_ALLOC_CACHED _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 1, 4)
#define CMA_ALLOC_NONCACHED _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 2, 4)
#define CMA_FREE _IOC(_IOC_WRITE, CMA_IOCTL_MAGIC, 3, 4)
#define CMA_GET_PHY_ADDR _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 4, 4)
#define CMA_GET_SIZE _IOC(_IOC_WRITE|_IOC_READ, CMA_IOCTL_MAGIC, 5, 4)
#define CMA_ALLOC_CACHED _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 1, 4)
#define CMA_ALLOC_NONCACHED _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 2, 4)
#define CMA_FREE _IOC(_IOC_WRITE, CMA_IOCTL_MAGIC, 3, 4)
#define CMA_GET_PHY_ADDR _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 4, 4)
#define CMA_GET_SIZE _IOC(_IOC_WRITE | _IOC_READ, CMA_IOCTL_MAGIC, 5, 4)

#define CMA_IOCTL_MAXNR 5
#define CMA_IOCTL_MAXNR 5

#ifndef CMA_DEBUG
#define CMA_DEBUG 0
#define CMA_DEBUG 0
#endif
#ifndef DRIVER_NODE_NAME
#define DRIVER_NODE_NAME "cma"
#define DRIVER_NODE_NAME "cma"
#endif

#if CMA_DEBUG == 1
#define __DEBUG(fmt, args...) printf("CMA_API_DEBUG: " fmt, ##args)
#define __DEBUG(fmt, args...) printf("CMA_API_DEBUG: " fmt, ##args)
#else
#define __DEBUG(fmt, args...)
#define __DEBUG(fmt, args...)
#endif

#define ROUND_UP(N, S) ((((N) + (S) - 1) / (S)) * (S))

#define ROUND_UP(N, S) ((((N) + (S)-1) / (S)) * (S))

/* Private functions */
void *cma_alloc(size_t size, unsigned ioctl_cmd);
void* cma_alloc(size_t size, unsigned ioctl_cmd);

/* Global file descriptor */
int cma_fd = 0;
Expand Down Expand Up @@ -99,23 +98,19 @@ int cma_release(void) {
return 0;
}

void *cma_alloc_cached(size_t size) {
return cma_alloc(size, CMA_ALLOC_CACHED);
}
void* cma_alloc_cached(size_t size) { return cma_alloc(size, CMA_ALLOC_CACHED); }

void *cma_alloc_noncached(size_t size) {
return cma_alloc(size, CMA_ALLOC_NONCACHED);
}
void* cma_alloc_noncached(size_t size) { return cma_alloc(size, CMA_ALLOC_NONCACHED); }

int cma_free(void *mem) {
int cma_free(void* mem) {
__DEBUG("Releasing contigous memory from 0x%x\n", (unsigned)mem);
unsigned data, v_addr;

/* save user space pointer value */
data = (unsigned)mem;
data = (unsigned)mem;
v_addr = (unsigned)mem;

if ( ioctl(cma_fd, CMA_GET_SIZE, &data) == -1 ) {
if (ioctl(cma_fd, CMA_GET_SIZE, &data) == -1) {
__DEBUG("cma_free - ioctl command unsuccsessful - 0\n");
return -1;
}
Expand All @@ -125,23 +120,23 @@ int cma_free(void *mem) {
munmap(mem, data);

/* free cma entry */
if ( ioctl(cma_fd, CMA_FREE, &v_addr) == -1 ) {
if (ioctl(cma_fd, CMA_FREE, &v_addr) == -1) {
__DEBUG("cma_free - ioctl command unsuccsessful - 1\n");
return -1;
}

return 0;
}

unsigned cma_get_phy_addr(void *mem) {
unsigned cma_get_phy_addr(void* mem) {
unsigned data;
__DEBUG("Getting physical address from 0x%x\n", (unsigned)mem);

/* save user space pointer value */
data = (unsigned)mem;

/* get physical address */
if ( ioctl(cma_fd, CMA_GET_PHY_ADDR, &data) == -1 ) {
if (ioctl(cma_fd, CMA_GET_PHY_ADDR, &data) == -1) {
__DEBUG("cma_free - ioctl command unsuccsessful\n");
return 0;
}
Expand All @@ -150,18 +145,17 @@ unsigned cma_get_phy_addr(void *mem) {
return data;
}


void *cma_alloc(size_t size, unsigned ioctl_cmd) {
void* cma_alloc(size_t size, unsigned ioctl_cmd) {
unsigned data;
void *mem;
void* mem;
__DEBUG("Allocating 0x%x bytes of contigous memory\n", size);

/* Page align size */
size = ROUND_UP(size, getpagesize());

/* ioctl cmd to allocate contigous memory */
data = (unsigned)size;
if ( ioctl(cma_fd, ioctl_cmd, &data) == -1 ) {
if (ioctl(cma_fd, ioctl_cmd, &data) == -1) {
__DEBUG("cma_alloc - ioctl command unsuccsessful\n");
return NULL;
}
Expand Down
82 changes: 52 additions & 30 deletions 3rdparty/compiler-rt/builtin_fp16.h
Original file line number Diff line number Diff line change
Expand Up @@ -29,16 +29,33 @@ static inline uint32_t __clz(uint32_t x) {
int n = 32;
uint32_t y;

y = x >>16; if (y) { n = n -16; x = y; }
y = x >> 8; if (y) { n = n - 8; x = y; }
y = x >> 4; if (y) { n = n - 4; x = y; }
y = x >> 2; if (y) { n = n - 2; x = y; }
y = x >> 1; if (y) return n - 2;
y = x >> 16;
if (y) {
n = n - 16;
x = y;
}
y = x >> 8;
if (y) {
n = n - 8;
x = y;
}
y = x >> 4;
if (y) {
n = n - 4;
x = y;
}
y = x >> 2;
if (y) {
n = n - 2;
x = y;
}
y = x >> 1;
if (y) return n - 2;
return n - x;
}

template <typename SRC_T, typename SRC_REP_T, int SRC_SIG_BITS,
typename DST_T, typename DST_REP_T, int DST_SIG_BITS>
template <typename SRC_T, typename SRC_REP_T, int SRC_SIG_BITS, typename DST_T, typename DST_REP_T,
int DST_SIG_BITS>
static inline DST_T __truncXfYf2__(SRC_T a) {
// Various constants whose values follow from the type parameters.
// Any reasonable optimizer will fold and propagate all of these.
Expand Down Expand Up @@ -71,7 +88,10 @@ static inline DST_T __truncXfYf2__(SRC_T a) {
const DST_REP_T dstNaNCode = dstQNaN - 1;

// Break a into a sign and representation of the absolute value
union SrcExchangeType { SRC_T f; SRC_REP_T i; };
union SrcExchangeType {
SRC_T f;
SRC_REP_T i;
};
SrcExchangeType src_rep;
src_rep.f = a;
const SRC_REP_T aRep = src_rep.i;
Expand All @@ -88,25 +108,21 @@ static inline DST_T __truncXfYf2__(SRC_T a) {

const SRC_REP_T roundBits = aAbs & roundMask;
// Round to nearest
if (roundBits > halfway)
absResult++;
// Ties to even
if (roundBits > halfway) absResult++;
// Ties to even
else if (roundBits == halfway)
absResult += absResult & 1;
}
else if (aAbs > srcInfinity) {
} else if (aAbs > srcInfinity) {
// a is NaN.
// Conjure the result by beginning with infinity, setting the qNaN
// bit and inserting the (truncated) trailing NaN field.
absResult = (DST_REP_T)dstInfExp << DST_SIG_BITS;
absResult |= dstQNaN;
absResult |= ((aAbs & srcNaNCode) >> (SRC_SIG_BITS - DST_SIG_BITS)) & dstNaNCode;
}
else if (aAbs >= overflow) {
} else if (aAbs >= overflow) {
// a overflows to infinity.
absResult = (DST_REP_T)dstInfExp << DST_SIG_BITS;
}
else {
} else {
// a underflows on conversion to the destination type or is an exact
// zero. The result may be a denormal or zero. Extract the exponent
// to get the shift amount for the denormalization.
Expand All @@ -124,24 +140,26 @@ static inline DST_T __truncXfYf2__(SRC_T a) {
absResult = denormalizedSignificand >> (SRC_SIG_BITS - DST_SIG_BITS);
const SRC_REP_T roundBits = denormalizedSignificand & roundMask;
// Round to nearest
if (roundBits > halfway)
absResult++;
// Ties to even
if (roundBits > halfway) absResult++;
// Ties to even
else if (roundBits == halfway)
absResult += absResult & 1;
}
}

// Apply the signbit to (DST_T)abs(a).
const DST_REP_T result = absResult | sign >> (srcBits - dstBits);
union DstExchangeType { DST_T f; DST_REP_T i; };
union DstExchangeType {
DST_T f;
DST_REP_T i;
};
DstExchangeType dst_rep;
dst_rep.i = result;
return dst_rep.f;
}

template<typename SRC_T, typename SRC_REP_T, int SRC_SIG_BITS,
typename DST_T, typename DST_REP_T, int DST_SIG_BITS>
template <typename SRC_T, typename SRC_REP_T, int SRC_SIG_BITS, typename DST_T, typename DST_REP_T,
int DST_SIG_BITS>
static inline DST_T __extendXfYf2__(SRC_T a) {
// Various constants whose values follow from the type parameters.
// Any reasonable optimizer will fold and propagate all of these.
Expand All @@ -157,15 +175,18 @@ static inline DST_T __extendXfYf2__(SRC_T a) {
const SRC_REP_T srcQNaN = SRC_REP_T(1) << (SRC_SIG_BITS - 1);
const SRC_REP_T srcNaNCode = srcQNaN - 1;

const int dstBits = sizeof(DST_T)*8;
const int dstBits = sizeof(DST_T) * 8;
const int dstExpBits = dstBits - DST_SIG_BITS - 1;
const int dstInfExp = (1 << dstExpBits) - 1;
const int dstExpBias = dstInfExp >> 1;

const DST_REP_T dstMinNormal = DST_REP_T(1) << DST_SIG_BITS;

// Break a into a sign and representation of the absolute value
union SrcExchangeType { SRC_T f; SRC_REP_T i; };
union SrcExchangeType {
SRC_T f;
SRC_REP_T i;
};
SrcExchangeType src_rep;
src_rep.f = a;
const SRC_REP_T aRep = src_rep.i;
Expand All @@ -191,8 +212,7 @@ static inline DST_T __extendXfYf2__(SRC_T a) {
absResult = (DST_REP_T)dstInfExp << DST_SIG_BITS;
absResult |= (DST_REP_T)(aAbs & srcQNaN) << (DST_SIG_BITS - SRC_SIG_BITS);
absResult |= (DST_REP_T)(aAbs & srcNaNCode) << (DST_SIG_BITS - SRC_SIG_BITS);
}
else if (aAbs) {
} else if (aAbs) {
// a is denormal.
// renormalize the significand and clear the leading bit, then insert
// the correct adjusted exponent in the destination type.
Expand All @@ -201,15 +221,17 @@ static inline DST_T __extendXfYf2__(SRC_T a) {
absResult ^= dstMinNormal;
const int resultExponent = dstExpBias - srcExpBias - scale + 1;
absResult |= (DST_REP_T)resultExponent << DST_SIG_BITS;
}
else {
} else {
// a is zero.
absResult = 0;
}

// Apply the signbit to (DST_T)abs(a).
const DST_REP_T result = absResult | (DST_REP_T)sign << (dstBits - srcBits);
union DstExchangeType { DST_T f; DST_REP_T i; };
union DstExchangeType {
DST_T f;
DST_REP_T i;
};
DstExchangeType dst_rep;
dst_rep.i = result;
return dst_rep.f;
Expand Down
Loading