Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Flakey Test] k8s.io/kubernetes/test/integration/client: metrics #127894

Closed
MadVikingGod opened this issue Oct 6, 2024 · 5 comments
Closed

[Flakey Test] k8s.io/kubernetes/test/integration/client: metrics #127894

MadVikingGod opened this issue Oct 6, 2024 · 5 comments
Labels
kind/flake Categorizes issue or PR as related to a flaky test. needs-sig Indicates an issue or PR lacks a `sig/foo` label and requires one. needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one.

Comments

@MadVikingGod
Copy link

Which jobs are flaking?

pull-kubernetes-integration

Which tests are flaking?

k8s.io/kubernetes/test/integration/client: metrics

There are others, but this is just a deep dive on this test.

Since when has it been flaking?

Unknown

Testgrid link

https://prow.k8s.io/view/gs/kubernetes-jenkins/pr-logs/pull/127819/pull-kubernetes-integration/1841824262885740544

Reason for failure (if possible)

This happens when the control plane leader election calls the OnStopLeading callback before the OnStartLeading callback, here.

A possible fix would be to fill the cancel function with an empty func() {} so it won't panic when this happens. I would be concerned that the run() properly handles an already canceled context.

Anything else we need to know?

I haven't been able to reproduce this locally, so I don't know if this fix will address this. I am willing to test this if anyone knows a way to get local failures.

Relevant SIG(s)

I don't know.

@MadVikingGod MadVikingGod added the kind/flake Categorizes issue or PR as related to a flaky test. label Oct 6, 2024
@k8s-ci-robot
Copy link
Contributor

There are no sig labels on this issue. Please add an appropriate label by using one of the following commands:

  • /sig <group-name>
  • /wg <group-name>
  • /committee <group-name>

Please see the group list for a listing of the SIGs, working groups, and committees available.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository.

@k8s-ci-robot k8s-ci-robot added the needs-sig Indicates an issue or PR lacks a `sig/foo` label and requires one. label Oct 6, 2024
@k8s-ci-robot
Copy link
Contributor

This issue is currently awaiting triage.

If a SIG or subproject determines this is a relevant issue, they will accept it by applying the triage/accepted label and provide further guidance.

The triage/accepted label can be added by org members by writing /triage accepted in a comment.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository.

@k8s-ci-robot k8s-ci-robot added the needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. label Oct 6, 2024
@googs1025
Copy link
Member

googs1025 commented Oct 7, 2024

hey @MadVikingGod
This has been temporarily resolved in this have #127834 (the fundamental problem still needs to be discussed FYI: #127737 (comment)). Have you git pulled the latest code?
/cc @aojea @Jefftree

@aojea
Copy link
Member

aojea commented Oct 7, 2024

/close

fixed by #127834

let's open the issues for flakes with a minimum of two occurrences, otherwise this does not scale , is not feasible to investigate any single failure and we need to correlate between at least two jobs to understand if is environmental or code related

@k8s-ci-robot
Copy link
Contributor

@aojea: Closing this issue.

In response to this:

/close

fixed by #127834

let's open the issues for flakes with a minimum of two occurrences, otherwise this does not scale , is not feasible to investigate any single failure and we need to correlate between at least two jobs to understand if is environmental or code related

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/flake Categorizes issue or PR as related to a flaky test. needs-sig Indicates an issue or PR lacks a `sig/foo` label and requires one. needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one.
Projects
None yet
Development

No branches or pull requests

4 participants