-
Notifications
You must be signed in to change notification settings - Fork 6.8k
[RFC] Large MXNet source files causing CI build failures #19688
Comments
There is a related issue: #18501 |
Thank you @mseth10 for looking into this !! Option1 may keep us blocked for long and has the uncertainty of how long will it take. IMHO Option 2 is the way forward :) |
+1 for option 2 |
A tangential question: Should we raise a bug report with |
How do you arrive at the time estimates? Adding a |
Effort will certainly vary depending on who does the work, their familiarity, and how long it takes them to ramp up. If we get people familiar with the code it could be shorter. If we get people unfamiliar, 5/day seems reasonable. If its someone new to MXNet (and possibly not an expert in C++) it could take longer. Until we get actual people to volunteer to do the work, and ask them how long they will need to do the work, we cant really give hard estimates. So these will have to do for now. |
A related problem is excessive code generation. Take That's:
That's 8 * 5 * 8 * 2 = 640 ways on CPU and 7 * 5 * 7 * 2 = 490 ways on GPU. This problem operates on a single axis. It reduces to: size of outer loop (i.e. the product of dimensions before the axis), the size of the axis in question, and the size of the data after the axis (i.e. the product of dimensions after the axis). After this simplification, there's no ndim dispatch. Supports arbitrary dimensionality with a factor of 5 reduction in compilation to 128 cases. In the common case where the types are the same and output is kWriteTo, a loop over memory copies is much faster. If we're just copying PODs, then the size of the data type can be folded into the size of the data to copy. So all 8 cases of identical input and output types with kWriteTo can be folded into one compilation, reducing 8 on CPU or 7 on GPU to 1. On CPU there are 121 cases: one for the normal copying operation and 120 for some combination of type conversion and/or kAddTo. On GPU there are 91 cases. |
To avoid build failures due to large source files. See #19688
Problem statement
MXNet CI is running OOM [1] while building MXNet binaries for unix-cpu and unix-gpu stages. This is an intermittent failure and the work around is to re-trigger CI a few times. The issue is caused due to some of the numpy .cc files being too large causing gcc to use too much memory. The issue was not pronounced with gcc7, but with the recent update to use gcc8 [2] for CI builds, we have started to see this OOM error.
The fix is to refactor the numpy .cc files into smaller files so that the objects created during compilation don't use much memory. Here is the list of the largest objects (>10MB in size) generated currently on Mac CPU build:
The corresponding cc files to above objects contains more than 210 operator registrations, and to refactor those into smaller files will need a considerable time and effort from the community. With 5 operators per day, that's more than 40 days of developers effort.
Proposed solutions
Option 1: We keep using gcc8 for CI builds and start working on refactoring these numpy .cc files. This would mean the community will have to face the CI failures for 40 days (could be less if more community members contribute).
Option 2: We go back to using gcc7 for CI builds, potentially solving the CI problem immediately, while we work on refactoring the numpy files. Reverting to gcc7 would take 2 days and then refactoring would take another 40 days.
I personally would prefer Option 2 for the reason that it saves contributors time in getting their PRs merged quickly, as well as saves on the CI resources. Would like to request community feedback on the same.
Going forward we also need to add a check to MXNet CI for build time memory usage. Any ideas for the same would be highly appreciated.
References
The text was updated successfully, but these errors were encountered: