Kubernetes master lost all configuration/reset after OOM
Bug #1816635 reported by
Tom Haddon
This bug affects 1 person
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Kubernetes Control Plane Charm |
Triaged
|
Medium
|
Unassigned |
Bug Description
About 11 hours ago, a relatively newly provisioned k8s cluster ran out of memory on the k8s-master (it was running in a VM with only 2GB of RAM and was the only master in the cluster). We can see a number of tracebacks in syslog, and the end result was that the cluster config was entirely reset - secrets that had been created disappeared, applications and pods that had been created disappeared.
Please let me know which logs you'd like to see to help figure out what the problem was.
This appears to be the first interesting log entry in syslog on the kubernetes-master from the time of the incident https:/
Changed in charm-kubernetes-master: | |
status: | Invalid → New |
summary: |
- Kubernetes master ran out of memory, and ended up resetting the cluster + Kubernetes master lost all configuration/reset |
Changed in charm-kubernetes-master: | |
assignee: | nobody → Mike Wilson (knobby) |
tags: | added: ci-regression-test |
Changed in charm-kubernetes-master: | |
assignee: | Mike Wilson (knobby) → nobody |
summary: |
- Kubernetes master lost all configuration/reset + Kubernetes master lost all configuration/reset after OOM |
Changed in charm-kubernetes-master: | |
importance: | Undecided → Critical |
status: | New → Triaged |
Changed in charm-kubernetes-master: | |
importance: | Critical → Medium |
To post a comment you must log in.
The official CDK bundles have a constraint on the master nodes for 4 gigs of memory. This is required to prevent the master node from running out of memory. I think this is just a case of not provisioning enough memory for the master.