-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Podman|CRI-O: unable to restart minikube once stopped #8033
Comments
Sounds like the cgroup manager again #7996 ? Or is there a subtle difference here ? |
It is still cgroup, but I think it's different as the runtime might (I don't know the minikube internals) act differently depending on the container runtime. |
Does PR #8001 help here ? Since you have issues starting the system container, I don't think it even gets as far as cri-o |
Actually I am testing minikube with your patch applied, otherwise as you're saying it won't start even the first time. |
Wonder why systemd worked in the previous podman container, but not in this one ? Or maybe that is what it didn't, and the same issue is showing here. Will try to reproduce. |
I've just downloaded the master branch again and, without applying the patch and having the same environment as this bug, the error is different. So yes, the patch seems to help, but it also seems that it's not complete. For reference, I'm showing you the log of the (re)start of minikube with the "vanilla" master branch:
|
Changing the host podman to use systemd/journald by default reproduces the issue.
|
Not sure why it didn't use Here is the unhelpful error message, when you do:
EDIT: Seemed to be a known error with podman 1.6.2, will try again. (should stop using that PPA) With podman 1.9.1, minikube goes back into waiting for the crashed container (with above error) |
Seems to be either a kernel or a compile option, neither is very helpful in this case ?
|
I've tried setting |
Btw the unhelpful error message might be integrated with the podman's log (as reported in the thread)
Strange thing is that it's happening only at the restart, not at the first start... |
It doesn't help that the path is missing or having duplicated components, either. This is the file created by start 1: And here is the bogus file in start 2: Basically a fight between two buggy pieces of software (podman and systemd) And the whole cgroups "v1" vs "v2" is not helping either, not a stable platform this. |
So actually someone, either podman or systemd, is missing the prefixed path, right? Is there any way we can try to debug this issue? Also in theory (AFAIK) both podman and CRI-O should support cgroup v2 out of the box (but not docker and containerd)... shouldn't the couple be running just fine with cgroup v2? (obviously this is not the case) |
As far as I know, it is podman that has an issue with cgroup manager=systemd and kubernetes that has an issue with cgroups v2 ? And docker and also runc, so might have to switch to crun as well... |
Maybe looking in the wrong place ? We do get the same error from docker too
But we don't get those other errors after the attach, that we get with podman:
|
This could be related: (but again, same for both drivers) |
I think it is as easy as the
Apparently |
Yes, that seems to have done the trick! (after start/stop) Custom entrypoint:
And now it can "allocate" the manager "object" again:
Basically the same code as for docker, for looping the cgroups. |
@afbjorklund is it possible to inject the podman support in the entrypoint somehow, or we have to edit the kubernete's image? I'm afraid I know too little to dig into this for a PR from my side... (was thinking about |
@elegos : the entrypoint comes from the KIND project, they don't have podman support (yet?). We should probably replace/patch in the image - eventually upstream - but for testing I just mounted it...
containers/podman#4373 (comment)
kubernetes-sigs/kind#154 (comment)
But this should at least fix one (additional) podman issue. Will make a kind PR for it, or if we fork the entrypoint (#7788) |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
Steps to reproduce the issue:
getenforce
(expected: "Permissive")systemd.unified_cgroup_hierarchy=0
), otherwise docker won't be able to run inside of podman.minikube start --driver=podman --conrtainer-runtime=cri-o
(expected OK)minikube stop
(expected OK)minikube start
(expected error)Full output of failed command:
Podman's minikube log
Searching on internet, I found out that balena OS had a very similar issue (see: https://forums.balena.io/t/container-reboot-and-failed-to-attach-1-to-compat-systemd-cgroup/27739/24). They suggested to resolve the issue running
systemd.legacy_systemd_cgroup_controller
before starting systemd (see solution in the linked issue).Thank you :)
The text was updated successfully, but these errors were encountered: