-
Notifications
You must be signed in to change notification settings - Fork 462
Setting kubelet log level to recommended defaults #2166
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Fedosin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/lgtm
|
Sounds great to me, every time I have to look at a node's journal the spam from hyperkube makes it much harder to read. Plus makes it more likely that useful information rotates out. So But I think you need to get this by the node team. |
|
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: cgwalters, Fedosin, kevchu3 The full list of commands accepted by this bot can be found here. The pull request process is described here DetailsNeeds approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
/retest |
|
there have definitely been times that we benefitted from having the default log level that high. I am leaning towards NACK on changing it. @rphillips @mrunalp @sjenning thoughts? |
|
Is the kubelet log level dynamically changeable via e.g. some grpc or something? How many cases would be covered if we made it easy for people to frob that on on one or more nodes? What about logging debug level things to a separate log file as kube-apiserver does? |
|
I think having it at 4 in CI is very helpful in chasing and fixing some of the races and bugs that we still have. In production, we could go down to a less verbose level. |
Hmmm. I suppose we could pretty easily teach our CI jobs to always inject a MachineConfig or kubeletconfig to bump this, but I think that's a somewhat dangerous path to go down because it makes our CI not look like production. That's a big trap. |
|
The logging volume level 4 generates into the infra indexes of the logging stack is significant in OCP 4.5. A more dynamic way to control the log level - globally or, on a per node basis for specific debugging, would better server both CI and users. |
|
I'm also in favor to make this change more dynamic so that one can choose what log-level is best. I can understand that in certain situation it will be good to have 4 but in normal production environments we may want to have 2. This is also based on experience where we have seen environments with about 70 nodes generating about 500 GB of infrastructure logs on daily basis, overloading the logging stack! |
|
/retest |
|
@kevchu3: The following tests failed, say
Full PR test history. Your PR dashboard. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. |
|
@kevchu3: PR needs rebase. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
|
superseded by: #2211 |
|
Is it recommend to have a verbosity --v =2 or 4 set for OpenShift Prod environment. We have an incident where ES disk was filling very fast and we are thinking of setting the verbosity to 2 for Prod. However a different point of view came in for Prod, as the goal is to minimize the time to reach a root cause conclusion as fast as possible and to avoid repeat issues. So, lowering the verbosity in prod and only increasing it after a failure or degradation event will require the event to re-occur quick because changing verbosity will cause the nodes affected by the verbosity change to be rebooted as part of restarting Kubelet process restart. This reset process could delay seeing the issue, especially if it's related to performance. Any thoughts or recommendations |
- What I did
The current kubelet log level is set to debug (
KUBELET_LOG_LEVEL=4), per this Red Hat Knowledgebase article 1. For default settings of a cluster, I'm proposing to set the log level to a more appropriate value (KUBELET_LOG_LEVEL=2). I've bolded the description of the proposed log level below:- How to verify it
Steps to verify are provided in this Knowledgebase article: https://access.redhat.com/solutions/4619431
- Description for the changelog
Setting kubelet log level to recommended defaults
Footnotes
https://access.redhat.com/solutions/4619431 ↩