Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Commit

Permalink
[mkldnn-v1.0] Skip flaky test for unidirectional rnn_relu
Browse files Browse the repository at this point in the history
Skip `test_rnnrelu_sym`, and add some issue tracking message
  • Loading branch information
zixuanweeei committed Oct 22, 2019
1 parent d109033 commit 5e40c7f
Showing 1 changed file with 3 additions and 0 deletions.
3 changes: 3 additions & 0 deletions tests/python/unittest/test_operator.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,7 @@
def check_rnn_consistency(cell1, cell2, T, N, I, H, grad_req, rtol=1e-2, atol=1e-4):
if default_context().device_type == 'cpu':
# NOTE(zixuanweeei): Currently, we don't add `add` requests support on fused mkl-dnn rnn operator.
# We tracked this issue by https://github.com/apache/incubator-mxnet/issues/16578
if isinstance(grad_req, dict) and 'add' in grad_req.values():
print("Skip the test when requiring `add` operation against gradients on CPU context.")
return
Expand Down Expand Up @@ -257,6 +258,8 @@ def test_rnntanh_bidirectional():
@with_seed()
@assert_raises_cudnn_not_satisfied(min_version='5.1.10')
def test_rnnrelu_sym():
if default_context().device_type == 'gpu':
print("Skip test `rnn_relu_sym` on gpu. This is tracked by https://github.com/apache/incubator-mxnet/issues/16548")
Ts = [1, 5]
Ns = [1, 32]
Is = [32, 128, 512]
Expand Down

0 comments on commit 5e40c7f

Please sign in to comment.