-
Notifications
You must be signed in to change notification settings - Fork 6.8k
fix custom exception handling #14575
Changes from 4 commits
da74a63
2b1da9e
e43f13d
5bce27e
23aee98
963be55
931aca9
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -96,7 +96,14 @@ class CustomOperator { | |
bool prev_recording = Imperative::Get()->set_is_recording(recording); | ||
bool prev_training = Imperative::Get()->set_is_training(training); | ||
|
||
func(); | ||
try { | ||
func(); | ||
} catch (dmlc::Error& e) { | ||
exception_ = | ||
std::make_shared<std::exception_ptr>(std::current_exception()); | ||
ctx.async_on_complete(); | ||
return; | ||
} | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I think we can solve both 1 and 2 this way: After func is called do wait_to_read on all elements in arrs. Then catch and save. Remove lines 104 and 105. In PushSync, check if exception is set and rethrow exception. Also catch it and call async_on_complete in pushsync. and return. Something like the following:
Thanks to this support added for horovod: #13932 we may be able to leverage this to call async_on_complete with the error. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Adding wait_to_read in custom op can solve 1&2, and it can be treated as normal op without using ExecType::kAsync. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. we probably still need PushSync for the Sparse ndarray updates. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. we still need ExecType::kAsync. Custom operator is still async and when push is called it just pushes it into its custom op worker queue for execution later. Async will ensure that the threaded_engine_pooled and threaded_engine_per_device treat it as a special case and execute immediately instead of pushing the work again to one of the engine worker thread queue. Pushing to engine worker thread queue is unnecessary for custom op. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. After testing, ExecType::kAsync is really needed. Adding wait_to_read in engine worker thread will cause deadlock. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. we probably still need it for sparse. since for sparse we are updating chunk it is a write option. WaitToRead may not be enough. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Yes, I also add WaitToWrite to make sure there's no left out exceptions. |
||
|
||
Imperative::Get()->set_is_training(prev_training); | ||
Imperative::Get()->set_is_recording(prev_recording); | ||
|
@@ -145,6 +152,7 @@ class CustomOperator { | |
num_free_threads = 0; | ||
destructing_ = false; | ||
naive_engine_ = true; | ||
exception_ = nullptr; | ||
if (std::string("NaiveEngine") != dmlc::GetEnv("MXNET_ENGINE_TYPE", std::string())) { | ||
naive_engine_ = false; | ||
} | ||
|
@@ -162,6 +170,14 @@ class CustomOperator { | |
workers_.clear(); | ||
} | ||
|
||
inline void ThrowException() { | ||
if (exception_ && *exception_) { | ||
std::exception_ptr tmp = *exception_; | ||
exception_ = nullptr; | ||
std::rethrow_exception(tmp); | ||
} | ||
} | ||
|
||
private: | ||
CustomOperator() { | ||
this->Start(); | ||
|
@@ -198,6 +214,7 @@ class CustomOperator { | |
std::vector<std::thread> workers_; | ||
std::atomic<uint32_t> num_free_threads; | ||
std::queue<std::function<void(void)> > q_; | ||
std::shared_ptr<std::exception_ptr> exception_; | ||
bool naive_engine_; | ||
bool destructing_; | ||
}; | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -29,6 +29,7 @@ | |
from mxnet.test_utils import * | ||
from mxnet.base import py_str, MXNetError, _as_list | ||
from common import setup_module, with_seed, teardown, assert_raises_cudnn_not_satisfied, assertRaises | ||
from nose.tools import assert_raises | ||
import unittest | ||
import os | ||
|
||
|
@@ -5200,29 +5201,29 @@ def create_operator(self, ctx, shapes, dtypes): | |
|
||
# test custom operator fork | ||
# see https://github.com/apache/incubator-mxnet/issues/14396 | ||
if not sys.platform.startswith('win'): # no fork in windows | ||
class AdditionOP(mx.operator.CustomOp): | ||
def __init__(self): | ||
super(AdditionOP, self).__init__() | ||
def forward(self, is_train, req, in_data, out_data, aux): | ||
out_data[0][:] = in_data[0] + in_data[1] | ||
def backward(self, req, out_grad, in_data, out_data, in_grad, aux): | ||
in_grad[0][:] = out_grad[0] | ||
in_grad[1][:] = out_grad[0] | ||
|
||
@mx.operator.register("AdditionOP") | ||
class AdditionOPProp(mx.operator.CustomOpProp): | ||
def __init__(self): | ||
super(AdditionOPProp, self).__init__() | ||
def list_arguments(self): | ||
return ['a', 'b'] | ||
def list_outputs(self): | ||
return ['output'] | ||
def infer_shape(self, in_shape): | ||
return in_shape, [in_shape[0]] | ||
def create_operator(self, ctx, shapes, dtypes): | ||
return AdditionOP() | ||
class AdditionOP(mx.operator.CustomOp): | ||
def __init__(self): | ||
super(AdditionOP, self).__init__() | ||
def forward(self, is_train, req, in_data, out_data, aux): | ||
out_data[0][:] = in_data[0] + in_data[1] | ||
def backward(self, req, out_grad, in_data, out_data, in_grad, aux): | ||
in_grad[0][:] = out_grad[0] | ||
in_grad[1][:] = out_grad[0] | ||
|
||
@mx.operator.register("AdditionOP") | ||
class AdditionOPProp(mx.operator.CustomOpProp): | ||
def __init__(self): | ||
super(AdditionOPProp, self).__init__() | ||
def list_arguments(self): | ||
return ['a', 'b'] | ||
def list_outputs(self): | ||
return ['output'] | ||
def infer_shape(self, in_shape): | ||
return in_shape, [in_shape[0]] | ||
def create_operator(self, ctx, shapes, dtypes): | ||
return AdditionOP() | ||
|
||
if not sys.platform.startswith('win'): # no fork in windows | ||
def custom_add(): | ||
a = mx.nd.array([1, 2, 3]) | ||
b = mx.nd.array([4, 5, 6]) | ||
|
@@ -5237,6 +5238,18 @@ def custom_add(): | |
p.join(5) | ||
assert not p.is_alive(), "deadlock may exist in custom operator" | ||
|
||
# test except handling | ||
# see https://github.com/apache/incubator-mxnet/pull/14575 | ||
def custom_add_exc(): | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Can we add a comment that an exception is expected due to shapes I assume? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Sure, I'll add it. |
||
a = mx.nd.array([1, 2, 3]) | ||
b = mx.nd.array([4, 5]) | ||
# trigger exception by providing unmatched operand shapes | ||
c = mx.nd.Custom(a, b, op_type='AdditionOP') | ||
c.wait_to_read() | ||
|
||
assert_raises(MXNetError, custom_add_exc) | ||
|
||
|
||
@with_seed() | ||
def test_psroipooling(): | ||
for num_rois in [1, 2]: | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Lets remove this. ThreadedEngine should not depend on custom operator code.