-
Notifications
You must be signed in to change notification settings - Fork 126
OCPBUGS-60546: UPSTREAM: 135680: test/e2e: increase memory limits in pod resize tests #2533
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
The pod resize e2e tests use memory limits as low as 20Mi for Guaranteed QoS pods. On OpenShift/CRI-O, the container runtime (runc) runs inside the pod's cgroup and requires ~20-22MB of memory during container creation and restart operations. This causes intermittent OOM kills when the pod's memory limit is at or below runc's memory footprint. This issue does not occur on containerd-based clusters because containerd's shim runs outside the pod's cgroup by default (ShimCgroup=""), so runc's memory is not charged against the pod's limit. Increase memory limits to provide sufficient headroom for runc: - originalMem: 20Mi -> 35Mi - reducedMem: 15Mi -> 30Mi - increasedMem: 25Mi -> 40Mi The test validates resize behavior, not minimal memory limits, so larger values do not reduce test coverage. Signed-off-by: Damien Grisonnet <[email protected]>
|
@dgrisonnet: This pull request references Jira Issue OCPBUGS-60546, which is valid. The bug has been moved to the POST state. 3 validation(s) were run on this bug
Requesting review from QA contact: The bug has been updated to refer to the pull request using the external bug tracker. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the openshift-eng/jira-lifecycle-plugin repository. |
|
@dgrisonnet: the contents of this pull request could not be automatically validated. The following commits could not be validated and must be approved by a top-level approver:
Comment |
|
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: dgrisonnet The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
/payload-aggregate periodic-ci-openshift-release-master-nightly-4.21-e2e-gcp-ovn-runc 10 |
|
@bitoku: trigger 1 job(s) for the /payload-(with-prs|job|aggregate|job-with-prs|aggregate-with-prs) command
See details on https://pr-payload-tests.ci.openshift.org/runs/ci/cb649f00-d504-11f0-8baf-64f82bef5ef1-0 |
|
@dgrisonnet: The following test failed, say
Full PR test history. Your PR dashboard. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. I understand the commands that are listed here. |
|
The job did flake once in the aggregate. |
|
/payload-aggregate periodic-ci-openshift-release-master-nightly-4.21-e2e-gcp-ovn-runc 10 |
|
@dgrisonnet: trigger 1 job(s) for the /payload-(with-prs|job|aggregate|job-with-prs|aggregate-with-prs) command
See details on https://pr-payload-tests.ci.openshift.org/runs/ci/ef076e00-d5af-11f0-8c43-a70d918b954d-0 |
Though it failed, this doesn't look OOM. |
|
/retest yeah |
|
/payload-aggregate periodic-ci-openshift-release-master-nightly-4.21-e2e-gcp-ovn-runc 20 |
|
@haircommander: trigger 1 job(s) for the /payload-(with-prs|job|aggregate|job-with-prs|aggregate-with-prs) command
See details on https://pr-payload-tests.ci.openshift.org/runs/ci/f2ee8f70-d600-11f0-8127-0592e288e31d-0 |
Pick kubernetes#135680 to fix downstream flakes with pod inplace resize tests.