Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Revert kube2sky from 1.2 back to 1.1 until we figure out why it's flaky. #7461

Merged
merged 1 commit into from
Apr 29, 2015

Conversation

cjcullen
Copy link
Member

No description provided.

@ghost
Copy link

ghost commented Apr 28, 2015

Have you run e2e tests against this to confirm that it reduces flakyness?

@cjcullen
Copy link
Member Author

1st round of e2e's passed. I'll run ~10 (up, Services-DNS-test, down) loops to see if I can coax any flakiness.

@zmerlynn
Copy link
Member

btw, doing up/test/down in a timely fashion is exactly what https://github.com/GoogleCloudPlatform/kubernetes/blob/master/hack/parallel-e2e.sh is for. :)

@a-robinson a-robinson assigned ghost Apr 28, 2015
@cjcullen
Copy link
Member Author

Between @zmerlynn and I, we've gotten ~15 up-DNS-down runs without a failure.

zmerlynn added a commit that referenced this pull request Apr 29, 2015
Revert kube2sky from 1.2 back to 1.1 until we figure out why it's flaky.
@zmerlynn zmerlynn merged commit 89195b0 into kubernetes:master Apr 29, 2015
@ghost
Copy link

ghost commented Apr 29, 2015

For what it's worth, I have run the tests 8 times locally now, without this fix, and none of them failed. So I'm not sure that ~15 local clean runs are a sure sign of improvement, but I'm happy to merge on the basis that it's unlikely to make things substantially worse.

@ghost
Copy link

ghost commented Apr 29, 2015

I'm not sure what version you were testing against in #7453 when you managed to reproduce the failing DNS test, but I just noticed that since kubernetes-ete-gce run 5340 at 11:17am today, the DNS e2e test has not failed once (i.e. even without the fix in this PR).

I couldn't help noticing that the following went into the above successful builds:

Commit 07400f9 by Wojciech Tyczynski
Increase maxIdleConnection limit in etcd client.

I wonder whether that fixed the problem? In which case we can potentially roll back this roll back.

@wojtek-t
Copy link
Member

I think that commit can help a lot for any connection-related issue to etcd (in particular it reduced cpu-usage for performance tests in apiserver by ~35%). But didn't dig into this issue, so I have no idea whether it was connection or performance related.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants