You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
So the Linear Scheduler for linearly increasing distortion (similar to dropblock) would in no way work for multi-gpu training since it uses a simple variable i (not tensor) so when we do the following
The value of i will never get updated. You can try if you want. My question is, how did you guys run this code to train imagenet and got those results?
The text was updated successfully, but these errors were encountered:
unless you're running it on a single gpu, the value of i will never be updated. It would remain 0 or change between 0 and 1 as it is not a tensor. It's a known issue and you can read more about it at,
In the code of disout, the update of i is not in the function 'forward()'. It is updated in function 'train' of file 'train_imagenet.py '. Thus the problem in the mentioned issue does not occur.
So the Linear Scheduler for linearly increasing distortion (similar to dropblock) would in no way work for multi-gpu training since it uses a simple variable i (not tensor) so when we do the following
The value of i will never get updated. You can try if you want. My question is, how did you guys run this code to train imagenet and got those results?
The text was updated successfully, but these errors were encountered: