-
Notifications
You must be signed in to change notification settings - Fork 712
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
failing e2e job kubeadm-kinder-kubelet-1-25-on-1-26 #2896
Comments
The sanboxImage of containerd is set as |
this error has been changed now, https://storage.googleapis.com/kubernetes-jenkins/logs/ci-kubernetes-e2e-kubeadm-kinder-kubelet-1-25-on-1-26/1670022676342640640/build-log.txt, the earliest err I can see now is
this might be caused by the missing the of the boot config file. Since this job is failed after the migration to the eks cluster, I'd suggest revert the change for this job to confirm this is an error caused by the cluster migration. |
edit: never mind, we never preloaded the sandbox imagevbased on containerd config. it should be really fast too, as it's <1mb it will be strange if the flakes are because of that. |
it seems we have to notify #sig-k8s-infra about the kernel config problem on eks nodes. @dims do you know who can help us? i am not so sure this is the cause of the failure, though. it is just a warning, and the nodes are the same for other test jobs too. but we can revert this test job to check for general eks node / kubelet incompatibility as you suggest. |
The revert PR was sent, feel free to review/merge it if we want to check if it's the reason of the CI failure. |
I didn't experience similar problems when testing locally. |
we do have more flakes after moving to eks, but perhaps we should only revert this problematic kubelet job? the flakes can be investigated separately. |
@SataQiu could we please have focused reverts? looking at this specific failure now. |
|
we do have a bump for |
and we have a daemonset ... dunno if that got deployed: |
ok got one green https://prow.k8s.io/?job=ci-kubernetes-e2e-kubeadm-kinder-kubelet-1-25-on-1-26 let's watch these jobs over the weekend and see if we spot other issues |
ok 3 greens in a row: https://prow.k8s.io/?job=ci-kubernetes-e2e-kubeadm-kinder-kubelet-1-25-on-1-26 |
thanks, @dims |
closing this; we can close the revert prs too, @SataQiu |
Great to hear that! For the record only, for the err msg: "inotify_init: too many open files", we need to open the |
https://testgrid.k8s.io/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-1-25-on-1-24
The last run. Not sure if this will be fixed by kubernetes/k8s.io#5438 We may wait for the next run. The start time of this CI is similar to the time that kubernetes/k8s.io#5438 was merged. |
https://testgrid.k8s.io/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-kubelet-1-25-on-1-26
i am seeing this in the kubelet logs:
started on 16.06
other kubelet jobs seem ok
The text was updated successfully, but these errors were encountered: