@@ -160,7 +160,7 @@ def __init__(
160
160
padding = (0 , input_kernel [1 ] // 2 , input_kernel [2 ] // 2 ) if self .padding else 0 ,
161
161
)
162
162
163
- self .add_bn_layer (layer = layer , hidden_channels = self . hidden_channels [ 0 ] )
163
+ self .add_bn_layer (layer = layer , layer_idx = 0 )
164
164
165
165
if layers > 1 or self .final_nonlinearity :
166
166
if hidden_nonlinearities == "adaptive_elu" :
@@ -185,7 +185,7 @@ def __init__(
185
185
padding = (0 , self .hidden_kernel [l ][1 ] // 2 , self .hidden_kernel [l ][2 ] // 2 ) if self .padding else 0 ,
186
186
)
187
187
188
- self .add_bn_layer (layer = layer , hidden_channels = self . hidden_channels [ l + 1 ] )
188
+ self .add_bn_layer (layer = layer , layer_idx = l + 1 )
189
189
190
190
if self .final_nonlinearity or l < self .layers :
191
191
if hidden_nonlinearities == "adaptive_elu" :
@@ -363,7 +363,10 @@ def __init__(
363
363
dilation = (self .temporal_dilation , 1 , 1 ),
364
364
)
365
365
366
- self .add_bn_layer (layer = layer , hidden_channels = self .hidden_channels [0 ])
366
+ self .add_bn_layer (
367
+ layer = layer ,
368
+ layer_idx = 0 ,
369
+ )
367
370
368
371
if layers > 1 or final_nonlin :
369
372
if hidden_nonlinearities == "adaptive_elu" :
@@ -394,7 +397,7 @@ def __init__(
394
397
dilation = (self .hidden_temporal_dilation [l ], 1 , 1 ),
395
398
)
396
399
397
- self .add_bn_layer (layer = layer , hidden_channels = self . hidden_channels [ l + 1 ] )
400
+ self .add_bn_layer (layer = layer , layer_idx = l + 1 )
398
401
399
402
if final_nonlin or l < self .layers :
400
403
if hidden_nonlinearities == "adaptive_elu" :
0 commit comments