Saturday, 2019 July 13
I pushed an update this morning (I'm trying to backfill some blog posts I've been woefully late in writing) and discovered that once again a coredns pod had gone south.
% kubectl describe endpoints --namespace=kube-system kube-dns
Name: kube-dns
Namespace: kube-system
Labels: k8s-app=kube-dns
kubernetes.io/cluster-service=true
kubernetes.io/name=CoreDNS
Annotations:
Subsets:
Addresses: 10.244.1.182
NotReadyAddresses: 10.244.0.16
Ports:
Name Port Protocol
---- ---- --------
dns 53 UDP
dns-tcp 53 TCP
metrics 9153 TCP
Events:
%
Deleting the bad pod (it's recreated automatically) "fixes" the problem for at least some length of time.
This is not something that used to happen, so I suspect some relatively recent update is triggering it. Now that I know what to look for I'm going to start paying much closer attention and see if I can track down the root cause.
Tuesday, 2019 July 09 Saturday, 2019 July 20