[Error] Context deadline exceeded on etcd v3.5.6 #15229
Unanswered
ajayudayagiri-hpe
asked this question in
Q&A
Replies: 2 comments 8 replies
-
Pretty old thread, but here's what you can try:
are you running anything special against your APIserver? I've seen kubeburner causing issues frequently (especially in namespace deletion scenarios). Or anything that attempts to list all pods, like a CNI plugin can be a culprit. |
Beta Was this translation helpful? Give feedback.
8 replies
-
Hello, I faced the same issue.
@jberkus @tjungblu @ajayudayagiri-hpe Is there any update for this? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
On a 5-node baremetal cluster we are continously seeing "context deadline exceeded" in etcd logs which is resulting in liveness probe failure of kube-apiserver. Also, the kube-apiserver pod is often restarting due to etcd unable to respond on time.
Below are the iterations we have performed to test this scenario.
Iteration 1 - Etcd as Pod
Iteration 2 - Etcd as System Service
The further setup configuration is similar in both iterations.
Setup Configuration
Cluster Size - 5
Master Nodes - 3
Member Nodes - 2
Kubernetes version - v1.22.17
Etcd version - v3.5.6
H/W of each node:
Log from etcd service
We have also used fio to check the disk I/O performance and the results are similar on all three nodes and are provided below. The parameters used below are --size=100m and --bs=2300.
As per above fio result the the 99th percentile of fsync data is around 265 or 0.26ms and is far below 10ms which is required for good performance of etcd. However, the context deadline exceeded issue is still seen.
Temporary fixes tried
On node reboot the context deadline exceeded issue goes away, however, it is seen again after few days and it continues to occur. It seems to be occurring in longevity.
Beta Was this translation helpful? Give feedback.
All reactions