Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[k8s.io] Dynamic provisioning [k8s.io] DynamicProvisioner should create and delete persistent volumes {Kubernetes e2e suite} #26682

Closed
k8s-github-robot opened this issue Jun 2, 2016 · 13 comments
Assignees
Labels
kind/flake Categorizes issue or PR as related to a flaky test. priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. sig/api-machinery Categorizes an issue or PR as relevant to SIG API Machinery.

Comments

@k8s-github-robot
Copy link

k8s-github-robot commented Jun 2, 2016

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke/8270/

Failed: [k8s.io] Dynamic provisioning [k8s.io] DynamicProvisioner should create and delete persistent volumes {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:123
Jun  1 22:01:48.966: Couldn't delete ns "e2e-tests-volume-provisioning-9pmxr": the server does not allow access to the requested resource (delete namespaces e2e-tests-volume-provisioning-9pmxr)
@caesarxuchao
Copy link
Member

This is a duplication of #26728, I'm looking into it.

@jsafrane
Copy link
Member

jsafrane commented Jun 3, 2016

This does not seem to be related to persistent volumes, the test is unable to delete a namespace:

Jun  1 22:01:48.966: INFO: Couldn't delete ns "e2e-tests-volume-provisioning-9pmxr": the server does not allow access to the requested resource (delete namespaces e2e-tests-volume-provisioning-9pmxr)

The namespace might contain a PersistentVolumeClaim or pods in various stages of termination, however it should be possible to delete it without errors.

Thanks @caesarxuchao for looking into it.

@matchstick matchstick added sig/api-machinery Categorizes an issue or PR as relevant to SIG API Machinery. and removed team/cluster labels Jun 7, 2016
@matchstick
Copy link
Contributor

After consulting with Chao we are moving this to team CSI due to namespace deletion.

@jsafrane jsafrane removed their assignment Jun 7, 2016
@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gce/18298/

Failed: [k8s.io] Dynamic provisioning [k8s.io] DynamicProvisioner should create and delete persistent volumes {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/volume_provisioning.go:123
Expected error:
    <*errors.errorString | 0xc820836090>: {
        s: "PersistentVolume pvc-ee236ef7-2cbe-11e6-9e3e-42010af00002 still exists within 20m0s",
    }
    PersistentVolume pvc-ee236ef7-2cbe-11e6-9e3e-42010af00002 still exists within 20m0s
not to have occurred

@jsafrane
Copy link
Member

jsafrane commented Jun 7, 2016

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gce/18298/

This is a different flake, volume controller can't delete provisioned GCE PD for 20 minutes:

I0607 15:00:16.246079       6 gce_util.go:58] Error deleting GCE PD volume jenkins-e2e-dynamic-pvc-ee236ef7-2cbe-11e6-9e3e-42010af00002: googleapi: Error 400: The disk resource 'jenkins-e2e-dynamic-pvc-ee236ef7-2cbe-11e6-9e3e-42010af00002' is already being used by 'jenkins-e2e-minion-group-zuq0', resourceInUseByAnotherResource

@saad-ali, is it a bug in #26351?

@saad-ali
Copy link
Member

saad-ali commented Jun 7, 2016

This is a different flake, volume controller can't delete provisioned GCE PD for 20 minutes:

Yes, 18298 hit a different issue than the original one being tracked by this bug. The persistent volume controller is unable to delete the provisioned volume because it is still attached after the test is complete. It remains attached because the attach_detach_controller apparently did not receive, or missed, the pod delete event.

#26406 is open to force the shared informer to resync. But resync wouldn't necessarily send out pod delete events, so it would be worth implementing a manual verifier that confirms each pod still exists periodically. Opened #26994 to track that.

@saad-ali
Copy link
Member

saad-ali commented Jun 8, 2016

because the attach_detach_controller apparently did not receive, or missed, the pod delete event

Missing delete event may be caused by #27004. Suggested work-around (#26994) should be implemented in addition to fixing that issue to prevent bugs in the informer from affecting the controller.

@k8s-github-robot
Copy link
Author

https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke/9472/

Failed: [k8s.io] Dynamic provisioning [k8s.io] DynamicProvisioner should create and delete persistent volumes {Kubernetes e2e suite}

/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/volume_provisioning.go:123
Expected error:
    <*errors.errorString | 0xc82056d6d0>: {
        s: "PersistentVolumeClaim pvc-6au2x not in phase Bound within 5m0s",
    }
    PersistentVolumeClaim pvc-6au2x not in phase Bound within 5m0s
not to have occurred

@k8s-github-robot k8s-github-robot added the priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. label Jun 19, 2016
@saad-ali
Copy link
Member

Looks like there are three different flavors of failure for this test:

  1. server does not allow access to the requested resource
    • Status: @matchstick--"After consulting with Chao we are moving this to team CSI due to namespace deletion."
  2. PersistentVolume pvc-ee236ef7-2cbe-11e6-9e3e-42010af00002 still exists within 20m0s
  3. PersistentVolumeClaim pvc-6au2x not in phase Bound within 5m0s

@jsafrane Could you take a look at 3 and see if the root cause is different from the others?

@jsafrane
Copy link
Member

@jsafrane Could you take a look at 3

I'd love to, but I can't find any kube-controller-manager.log

@caesarxuchao
Copy link
Member

#27004 should be fixed by #27435

@caesarxuchao
Copy link
Member

but I can't find any kube-controller-manager.log

GKE had logging issues, which is fixed now. We can close this issue and wait for new occurrences. Is there any objection?

@saad-ali
Copy link
Member

saad-ali commented Jul 7, 2016

No objection. Closing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/flake Categorizes issue or PR as related to a flaky test. priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. sig/api-machinery Categorizes an issue or PR as relevant to SIG API Machinery.
Projects
None yet
Development

No branches or pull requests

6 participants