Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

[v1.x] Backport #17702 and #17872 to v1.x branch #18038

Merged
merged 2 commits into from
Apr 15, 2020

Conversation

zixuanweeei
Copy link
Contributor

Description

As title. #17702 and #17872 revised same lines in test_gluon_rnn.py. So we need to backport them in one. @ciyongch @TaoLv @pengzhao-intel

Also cc @stu1130.

…7702)

* Support projection feature for LSTM on CPU

* test solution for -Werror=maybe-uninitialized

* Check device type when create state

* Document the projection feature of LSTM for RNN operator

* Minor fix

* Re-run CI
…che#17872)

* Fix issue of zeros gradients w.r.t. RNN bias when num_layers > 1

* Use nd.copy() to initialize parameters of new operator

* Add check for output states

* Initialize i2h/h2h_weights with zeros for rnn_relu/tanh, and reduce size

* Split fused rnn layer test into tests of individual mode

* Skip lstm and gru tests on CPU context without DNNL
@zixuanweeei zixuanweeei requested a review from szha as a code owner April 13, 2020 01:58
@mxnet-bot
Copy link

Hey @zixuanweeei , Thanks for submitting the PR
All tests are already queued to run once. If tests fail, you can trigger one or more tests again with the following commands:

  • To trigger all jobs: @mxnet-bot run ci [all]
  • To trigger specific jobs: @mxnet-bot run ci [job1, job2]

CI supported jobs: [clang, edge, unix-gpu, unix-cpu, centos-cpu, windows-cpu, centos-gpu, windows-gpu, website, miscellaneous, sanity]


Note:
Only following 3 categories can trigger CI :PR Author, MXNet Committer, Jenkins Admin.
All CI tests must pass before the PR can be merged.

@ciyongch
Copy link
Contributor

Thanks @zixuanweeei, adding this to 1.7.0 roadmap #16864

@zixuanweeei
Copy link
Contributor Author

@mxnet-bot run ci [centos-gpu, unix-gpu]

@mxnet-bot
Copy link

Jenkins CI successfully triggered : [centos-gpu, unix-gpu]

@mxnet-bot
Copy link

Unauthorized access detected.
Only following 3 categories can trigger CI :
PR Author, MXNet Committer, Jenkins Admin.

@stu1130
Copy link
Contributor

stu1130 commented Apr 13, 2020

@zixuanweeei Thanks.

I saw the PR is for branch v1.x. Would v1.7x have this PR?

@leezu
Copy link
Contributor

leezu commented Apr 13, 2020

[2020-04-13T18:39:37.116Z] W: Failed to fetch https://developer.download.nvidia.com/compute/cuda/repos/ubuntu1404/x86_64/Packages  gnutls_handshake() failed: Handshake failed

[2020-04-13T18:39:37.116Z] 

[2020-04-13T18:39:37.116Z] W: Failed to fetch https://developer.download.nvidia.com/compute/machine-learning/repos/ubuntu1404/x86_64/Packages  gnutls_handshake() failed: Handshake failed

Please include Chai's recent fix: #18018

@leezu
Copy link
Contributor

leezu commented Apr 13, 2020

@stu1130 all 1.x commits made prior to a set date will be included in 1.7.

@ChaiBapchya
Copy link
Contributor

Here : #18044

@zixuanweeei
Copy link
Contributor Author

@mxnet-bot run ci [unix-gpu]

@mxnet-bot
Copy link

Jenkins CI successfully triggered : [unix-gpu]

Copy link
Contributor

@stu1130 stu1130 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@pengzhao-intel
Copy link
Contributor

Merging now.

@pengzhao-intel pengzhao-intel merged commit 6fa374b into apache:v1.x Apr 15, 2020
stu1130 pushed a commit to stu1130/incubator-mxnet that referenced this pull request Apr 15, 2020
…18038)

* Support projection feature for LSTM on CPU (Only Inference) (apache#17702)

* Support projection feature for LSTM on CPU

* test solution for -Werror=maybe-uninitialized

* Check device type when create state

* Document the projection feature of LSTM for RNN operator

* Minor fix

* Re-run CI

* Fix issue of zeros gradients w.r.t. RNN bias when num_layers > 1 (apache#17872)

* Fix issue of zeros gradients w.r.t. RNN bias when num_layers > 1

* Use nd.copy() to initialize parameters of new operator

* Add check for output states

* Initialize i2h/h2h_weights with zeros for rnn_relu/tanh, and reduce size

* Split fused rnn layer test into tests of individual mode

* Skip lstm and gru tests on CPU context without DNNL
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

7 participants