-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AAW Dev: Re-size workloads scheduled on system nodepools #1992
Comments
I created excel tables of cpu and memory utilization averages of all pods on the system nodes in aaw-dev. |
Here are the tables for cpu and memory usage of pods running on system nodes and suggested requests: |
Currently tracking down the parent objects and manifests corresponding to all the pods listed in the tables. |
CPU-wise the sum of all the CPU requests in the column goes up to 3.48 vCPU, this would fit on two system nodes sized to using 2 D2s, still want to better size the requests of course, but I will note an inaccuracy in the Also need to be careful of making sure the memory requests are honoured as well, as if we move to a D2 we get 8Gb of memory |
Took a bit of time to go over the pods listed in Jacek's excel file and made a few notes; Daemonsets you can easily influencestatcan-system/sysctl, azure-blob-csi-system/csi-blob-node: both of these are already well-sized already though Daemonsets deployed via helmThese ones I'm not too sure on as some may just be deployed by CNS Daemonsets with no traceable (might just be CNS / AKS)azure-ip-masq-agent, azure-npm, cloud-node-manager, csi-azuredisk-node, csi-azurefile-node, istio-cni-node, kube-proxy Deployments that you don't need to change (no requests)cert-manager-anything, kube-prometheus-stack-kube-state-metrics, statcan-system/toleration-injector (only has limits) Likely dont have the ability to changecoredns(100m,70Mi), coredns-autoscaler(20m,10Mi), konnectivity-agent(20,20), metrics-server(5m,30Mi), Resources deployed by helmgatekeeper-audit(100m, 1546Mi), gatekeeper-controller-manager(100m, 1546Mi), Argo resourcesstatcan-system/sidecar-terminator(10m,200M) |
My discovery work so far. Identified four argocd managed workloads and a bunch of aks managed workloads. Not sure about the helm managed ones yet. |
Take the pod workloads that you can see on the
system
workloads and after consulting with grafana over an extended period of time (the option) suggest and change workload sizes.It would be nice to have a table kind of like this
Follow up issue for general nodepool #1997
The text was updated successfully, but these errors were encountered: