-
Notifications
You must be signed in to change notification settings - Fork 4.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error removing intermediate container 20dbfe8b8d9d: Driver devicemapper failed to remove root filesystem #12923
Comments
Docker version on that test is:
We have not bumped it in a while (we have had it locked to |
@stevekuznetsov There is a new docker build coming that will add the ID of the device that cannot be deleted. Using that ID and a script from https://bugzilla.redhat.com/show_bug.cgi?id=1391665 we will be able to determine which application is holding that device busy during the attempted delete. |
@jwhonce these failures happen on ephemeral test VMs so it will be impossible for someone to come in and run the script. Are you suggesting that we need to create an exit check for the |
We actually got our single-node cluster (ha!) stuck with this issue since a couple of days (probably due to some upgrade or sheer coincidence) and we are not able to some builds Any help on how to debug this issue is more than welcome I've tried running the script ex-post but I always get "No Pid" output |
A bit of differential analysis between a build "type" that succeeds and one that fails if it is of any help:
It could be that either nodejs, npm or webpack fail to release some kind of resource which in turn creates issues to docker or openshift. |
@nanomad You are able to reproduce the failure consistently with the build that uses |
@stevekuznetsov We can even go a step further and provide ssh access to RH / Openshift developers if needed. The machine is not a PROD environment to begin with. If you would like to do so, just send me a message on GitHub. The installation details are the following:
|
Cool. @jwhonce would be interested in SSH I believe, thank you for your help. The reproducer is huge -- I don't think we had one before. |
@stevekuznetsov The more failures we can capture the faster we can zero in on the root cause. If it's possible to automate the capture even better! Thanks. |
@stevekuznetsov , @jwhonce you should receive a mail from me shortly |
@jwhonce - Should this enter the same MODIFIED/believed-to-be-fixed state as https://bugzilla.redhat.com/show_bug.cgi?id=1391665 ? (Or are we uncertain if this is tracking the same issue.) cc: @stevekuznetsov |
Re-assigned to @bparees as he provided fix to bug |
we still see these issues just during normal docker builds when docker is trying to clean up intermediate containers. however they do not appear to cause failures. i'll close this for now(since the failure mode should be fixed by the s2i changes to allow the commit to take longer) and open a new issue for the docker build errors we're seeing. |
as seen in:
https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_future/93/consoleFull#-62719347577f0ce7e4b0b14b5836ce6d
we previously had #9548 and #9490 in this space, but they got pretty messy and ultimately closed, but it's not clear to me if we think it should be working at this point or not (or maybe we have a bad docker in our AMIs).... so i'm starting the conversation here. Is the situation that:
our AMI used to contain the fix and our AMI regressed for some reason?
@stevekuznetsov @jwhonce @runcom
The text was updated successfully, but these errors were encountered: