-
Notifications
You must be signed in to change notification settings - Fork 40k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failing SIG-Node presubmit jobs #127831
Comments
we'd at least ensure all the blocking presubmit are healthy, and we can probably clean up the older ones if nobody cares about them |
Generally, most presubmit jobs has another periodic job.
Hence, most of the failure may be tracked in the periodic job flake or failing test.
Those will be tracked by release signal team as the failure/flake can also be found in https://testgrid.k8s.io/sig-release-master-blocking. I suppose, we don't need to track these two here. |
This is tracked in #127312. for
|
After kubernetes/test-infra#33633, |
flaky(known flaking) and canary tests are with lower priority.
This should be fixed by #127495.
See #127996.
@bart0sh your fix kubernetes/test-infra#33640 was merged today. It looks green now.
https://testgrid.k8s.io/sig-node-kubelet#kubelet-gce-e2e-arm64-ubuntu-serial looks green.
Add to my todo list. (will update here once do a diff): kubernetes/test-infra#33641
Added recently. See #127484. I plan to open an issue to track it. #128042 was opened
Related to #127312. I prefer to think they are related to the Evented PLEG issue.
https://testgrid.k8s.io/sig-node-release-blocking#node-kubelet-containerd-standalone-mode-all-alpha
https://testgrid.k8s.io/sig-node-presubmits#pr-node-kubelet-containerd-alpha-features
There are some flakes and the above one is always failing. |
@pacoxu Thank you for the detailed overview and the links to the issues! I'll try to continue investigating (and hopefully fixing) pr jobs and running them in the test PR. |
@pacoxu I excluded -kubetest2 jobs from the scope of this issue as they seem to be a work in progress. |
Do you mean the migration process kubernetes/test-infra#32567? |
Yes. I decided to concentrate on more stable jobs. I'm hoping that at the end of this road the -kubetest2 suffix will be removed and my test PR will trigger them automatically. They're still too buggy to pay attention to them right now in my opinion. |
/triage accepted TY for looking into this! |
@bart0sh should you be assigned to this? |
@kannon92 thanks for the reminder! |
|
Just see https://testgrid.k8s.io/presubmits-kubernetes-blocking this is a blocking board for presubmits CIs.
|
@pacoxu @kannon92 @haircommander I'm going to close this issue as almost all pr jobs failures have been fixed. There are only 4 failing jobs (see my test pr) currently:
Any thoughts/objections/suggestions? |
Thank you for the work! Happy to close |
Thank you! Closing. |
Great job. Thanks @bart0sh. |
Thank you it's been a great run! The next step would be to develop a setup which makes sure that no pr job is forgotten. Otherwise we'll end up with the same situation after some time. I'd propose to make sure every pr job is mapped to the ci job with the same name, e.g. Any other ideas? |
Which jobs are failing?
Which tests are failing?
Since when has it been failing?
I believe that most of the jobs were not triggered for a long time, so it's hard to say for how long they're failing.
Testgrid links
https://testgrid.k8s.io/sig-node-presubmits
https://testgrid.k8s.io/sig-node-ec2
Reason for failure (if possible)
I've triggered all SIG-Node pull* jobs for my test PR (the codebase is the same as the latest master branch). Here is how the list of jobs was generated:
Anything else we need to know?
No response
Relevant SIG(s)
/sig node
The text was updated successfully, but these errors were encountered: