diff --git a/python/mxnet/gluon/contrib/nn/basic_layers.py b/python/mxnet/gluon/contrib/nn/basic_layers.py index ebe136e30208..6cbf988fc94a 100644 --- a/python/mxnet/gluon/contrib/nn/basic_layers.py +++ b/python/mxnet/gluon/contrib/nn/basic_layers.py @@ -199,10 +199,10 @@ class SyncBatchNorm(BatchNorm): Initializer for the beta weight. gamma_initializer: str or `Initializer`, default 'ones' Initializer for the gamma weight. - moving_mean_initializer: str or `Initializer`, default 'zeros' - Initializer for the moving mean. - moving_variance_initializer: str or `Initializer`, default 'ones' - Initializer for the moving variance. + running_mean_initializer: str or `Initializer`, default 'zeros' + Initializer for the running mean. + running_variance_initializer: str or `Initializer`, default 'ones' + Initializer for the running variance. Inputs: diff --git a/python/mxnet/gluon/nn/basic_layers.py b/python/mxnet/gluon/nn/basic_layers.py index f8566dd05aa5..3d6976c32740 100644 --- a/python/mxnet/gluon/nn/basic_layers.py +++ b/python/mxnet/gluon/nn/basic_layers.py @@ -305,10 +305,10 @@ class BatchNorm(HybridBlock): Initializer for the beta weight. gamma_initializer: str or `Initializer`, default 'ones' Initializer for the gamma weight. - moving_mean_initializer: str or `Initializer`, default 'zeros' - Initializer for the moving mean. - moving_variance_initializer: str or `Initializer`, default 'ones' - Initializer for the moving variance. + running_mean_initializer: str or `Initializer`, default 'zeros' + Initializer for the running mean. + running_variance_initializer: str or `Initializer`, default 'ones' + Initializer for the running variance. in_channels : int, default 0 Number of channels (feature maps) in input data. If not specified, initialization will be deferred to the first time `forward` is called