Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Flaky test: sgld #14241

Closed
eric-haibin-lin opened this issue Feb 23, 2019 · 2 comments · Fixed by #14473
Closed

Flaky test: sgld #14241

eric-haibin-lin opened this issue Feb 23, 2019 · 2 comments · Fixed by #14473

Comments

@eric-haibin-lin
Copy link
Member

======================================================================

FAIL: test_operator_gpu.test_sgld

----------------------------------------------------------------------

Traceback (most recent call last):

  File "/usr/local/lib/python3.5/dist-packages/nose/case.py", line 198, in runTest

    self.test(*self.arg)

  File "/usr/local/lib/python3.5/dist-packages/nose/util.py", line 620, in newfunc

    return func(*arg, **kw)

  File "/work/mxnet/tests/python/gpu/../unittest/common.py", line 173, in test_new

    orig_test(*args, **kwargs)

  File "/work/mxnet/tests/python/gpu/../unittest/test_optimizer.py", line 520, in test_sgld

    compare_optimizer_noise_seeded(opt1(**kwarg), opt2(**kwarg), shape, dtype, seed)

  File "/work/mxnet/tests/python/gpu/../unittest/test_optimizer.py", line 511, in compare_optimizer_noise_seeded

    assert_almost_equal(w1.asnumpy(), w2.asnumpy(), rtol=rtol, atol=atol)

  File "/work/mxnet/python/mxnet/test_utils.py", line 495, in assert_almost_equal

    raise AssertionError(msg)

AssertionError: 

Items are not equal:

Error 5.851562 exceeds tolerance rtol=0.000100, atol=0.000010.  Location of maximum error:(2, 2, 0), a=0.734863, b=0.734375

 a: array([[[ 0.628   ,  0.6836  ,  0.2057  ,  0.909   ,  0.3188  ],

        [ 0.2113  ,  0.3286  ,  0.8438  ,  0.847   ,  0.5557  ],

        [ 0.6147  ,  0.595   ,  0.858   ,  0.0252  ,  0.1906  ],...

 b: array([[[ 0.628   ,  0.6836  ,  0.2057  ,  0.909   ,  0.3188  ],

        [ 0.2113  ,  0.3286  ,  0.8438  ,  0.847   ,  0.5557  ],

        [ 0.6147  ,  0.595   ,  0.858   ,  0.0252  ,  0.1906  ],...

-------------------- >> begin captured logging << --------------------

common: INFO: Setting test np/mx/python random seeds, use MXNET_TEST_SEED=1061286115 to reproduce.

--------------------- >> end captured logging << ---------------------

occured in http://jenkins.mxnet-ci.amazon-ml.com/blue/organizations/jenkins/mxnet-validation%2Funix-gpu/activity for PR #14237

@mxnet-label-bot
Copy link
Contributor

Hey, this is the MXNet Label Bot.
Thank you for submitting the issue! I will try and suggest some labels so that the appropriate MXNet community members can help resolve it.
Here are my recommended labels: Test, Flaky

@frankfliu
Copy link
Contributor

@mxnet-label-bot add [Flaky, test]

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

Successfully merging a pull request may close this issue.

4 participants